SOTAVerified

GPU

Papers

Showing 676700 of 5629 papers

TitleStatusHype
Act Now: A Novel Online Forecasting Framework for Large-Scale Streaming DataCode1
Global Tensor Motion PlanningCode1
ADAF: An Artificial Intelligence Data Assimilation Framework for Weather ForecastingCode1
Quantization without TearsCode1
ITER: Iterative Transformer-based Entity Recognition and Relation ExtractionCode1
GPU-Accelerated Inverse Lithography Towards High Quality Curvy Mask GenerationCode1
Diffusion Sampling Correction via Approximately 10 ParametersCode1
HRDecoder: High-Resolution Decoder Network for Fundus Image Lesion SegmentationCode1
LiVOS: Light Video Object Segmentation with Gated Linear MatchingCode1
Fast and Memory-Efficient Video Diffusion Using Streamlined InferenceCode1
KD-LoRA: A Hybrid Approach to Efficient Fine-Tuning with LoRA and Knowledge DistillationCode1
LOGO -- Long cOntext aliGnment via efficient preference OptimizationCode1
KVSharer: Efficient Inference via Layer-Wise Dissimilar KV Cache SharingCode1
syren-new: Precise formulae for the linear and nonlinear matter power spectra with massive neutrinos and dynamical dark energyCode1
xPerT: Extended Persistence TransformerCode1
EP-SAM: Weakly Supervised Histopathology Segmentation via Enhanced Prompt with Segment AnythingCode1
SPA: 3D Spatial-Awareness Enables Effective Embodied RepresentationCode1
Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual ExplanationsCode1
PostEdit: Posterior Sampling for Efficient Zero-Shot Image EditingCode1
Large Language Model Inference Acceleration: A Comprehensive Hardware PerspectiveCode1
LLM-Pilot: Characterize and Optimize Performance of your LLM Inference ServicesCode1
Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads on Consumer-Grade DevicesCode1
TorchSISSO: A PyTorch-Based Implementation of the Sure Independence Screening and Sparsifying Operator for Efficient and Interpretable Model DiscoveryCode1
STGformer: Efficient Spatiotemporal Graph Transformer for Traffic ForecastingCode1
Analog In-Memory Computing Attention Mechanism for Fast and Energy-Efficient Large Language ModelsCode1
Show:102550
← PrevPage 28 of 226Next →

No leaderboard results yet.