SOTAVerified

GPU

Papers

Showing 276300 of 5629 papers

TitleStatusHype
MMInference: Accelerating Pre-filling for Long-Context VLMs via Modality-Aware Permutation Sparse AttentionCode5
Hexcute: A Tile-based Programming Language with Automatic Layout and Task-Mapping Synthesis0
Scalable APT Malware Classification via Parallel Feature Extraction and GPU-Accelerated Learning0
A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained SettingsCode0
Splitwiser: Efficient LM inference with constrained resourcesCode0
LithOS: An Operating System for Efficient Machine Learning on GPUs0
Distribution-aware Dataset Distillation for Efficient Image Restoration0
Robust and Real-time Surface Normal Estimation from Stereo Disparities using Affine Transformations0
Beyond Terabit/s Integrated Neuromorphic Photonic Processor for DSP-Free Optical Interconnects0
SG-Reg: Generalizable and Efficient Scene Graph RegistrationCode2
AlphaZero-Edu: Making AlphaZero Accessible to EveryoneCode0
HPU: High-Bandwidth Processing Unit for Scalable, Cost-effective LLM Inference via GPU Co-processing0
Quantum Walks-Based Adaptive Distribution Generation with Efficient CUDA-Q Acceleration0
Lightweight LiDAR-Camera 3D Dynamic Object Detection and Multi-Class Trajectory PredictionCode1
NNTile: a machine learning framework capable of training extremely large GPT language models on a single node0
Mask Image WatermarkingCode1
Second-order Optimization of Gaussian Splats with Importance Sampling0
ARAP-GS: Drag-driven As-Rigid-As-Possible 3D Gaussian Splatting Editing with Diffusion Prior0
Tilus: A Virtual Machine for Arbitrary Low-Precision GPGPU Computation in LLM Serving0
Data-efficient LLM Fine-tuning for Code GenerationCode1
Taccel: Scaling Up Vision-based Tactile Robotics via High-performance GPU SimulationCode2
BitNet b1.58 2B4T Technical Report0
Accelerating Clinical NLP at Scale with a Hybrid Framework with Reduced GPU Demands: A Case Study in Dementia Identification0
MOM: Memory-Efficient Offloaded Mini-Sequence Inference for Long Context Language Models0
Characterizing and Optimizing LLM Inference Workloads on CPU-GPU Coupled Architectures0
Show:102550
← PrevPage 12 of 226Next →

No leaderboard results yet.