| ClusComp: A Simple Paradigm for Model Compression and Efficient Finetuning | Mar 17, 2025 | GPUModel Compression | —Unverified | 0 |
| Long-VMNet: Accelerating Long-Form Video Understanding via Fixed Memory | Mar 17, 2025 | FormGPU | —Unverified | 0 |
| MagicDistillation: Weak-to-Strong Video Distillation for Large-Scale Few-Step Synthesis | Mar 17, 2025 | GPU | —Unverified | 0 |
| AccelGen: Heterogeneous SLO-Guaranteed High-Throughput LLM Inference Serving for Diverse Applications | Mar 17, 2025 | ChunkingGPU | —Unverified | 0 |
| Changing Base Without Losing Pace: A GPU-Efficient Alternative to MatMul in DNNs | Mar 15, 2025 | GPU | —Unverified | 0 |
| PIPO: Pipelined Offloading for Efficient Inference on Consumer Devices | Mar 15, 2025 | GPUScheduling | —Unverified | 0 |
| Characterizing GPU Resilience and Impact on AI/HPC Systems | Mar 14, 2025 | AttributeGPU | —Unverified | 0 |
| Vamba: Understanding Hour-Long Videos with Hybrid Mamba-Transformers | Mar 14, 2025 | GPUMamba | —Unverified | 0 |
| Distance-Based Tree-Sliced Wasserstein Distance | Mar 14, 2025 | Computational EfficiencyGPU | CodeCode Available | 0 |
| X-EcoMLA: Upcycling Pre-Trained Attention into MLA for Efficient and Extreme KV Compression | Mar 14, 2025 | GPU | —Unverified | 0 |
| LLMPerf: GPU Performance Modeling meets Large Language Models | Mar 14, 2025 | GPU | CodeCode Available | 0 |
| Cost-effective Deep Learning Infrastructure with NVIDIA GPU | Mar 14, 2025 | Deep LearningGPU | CodeCode Available | 0 |
| OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models | Mar 13, 2025 | channel selectionContrastive Learning | —Unverified | 0 |
| KV-Distill: Nearly Lossless Learnable Context Compression for LLMs | Mar 13, 2025 | GPUQuestion Answering | —Unverified | 0 |
| Speedy MASt3R | Mar 13, 2025 | 3D Scene ReconstructionGPU | —Unverified | 0 |
| MoE-Gen: High-Throughput MoE Inference on a Single GPU with Module-Based Batching | Mar 12, 2025 | GPU | CodeCode Available | 0 |
| Priority-Aware Preemptive Scheduling for Mixed-Priority Workloads in MoE Inference | Mar 12, 2025 | BlockingGPU | —Unverified | 0 |
| Sometimes Painful but Certainly Promising: Feasibility and Trade-offs of Language Model Inference at the Edge | Mar 12, 2025 | CPUGPU | —Unverified | 0 |
| VideoScan: Enabling Efficient Streaming Video Understanding via Frame-level Semantic Carriers | Mar 12, 2025 | GPUStreaming video understanding | —Unverified | 0 |
| MarineGym: A High-Performance Reinforcement Learning Platform for Underwater Robotics | Mar 12, 2025 | BenchmarkingGPU | —Unverified | 0 |
| Mind the Memory Gap: Unveiling GPU Bottlenecks in Large-Batch LLM Inference | Mar 11, 2025 | GPU | CodeCode Available | 0 |
| Accelerating MoE Model Inference with Expert Sharding | Mar 11, 2025 | DecoderGPU | —Unverified | 0 |
| TT-GaussOcc: Test-Time Compute for Self-Supervised Occupancy Prediction via Spatio-Temporal Gaussian Splatting | Mar 11, 2025 | GPU | —Unverified | 0 |
| AdaptSR: Low-Rank Adaptation for Efficient and Scalable Real-World Super-Resolution | Mar 10, 2025 | GPUSuper-Resolution | —Unverified | 0 |
| Global Context Is All You Need for Parallel Efficient Tractography Parcellation | Mar 10, 2025 | AllData Augmentation | —Unverified | 0 |