SOTAVerified

GPU

Papers

Showing 251300 of 5629 papers

TitleStatusHype
MobileVLM : A Fast, Strong and Open Vision Language Assistant for Mobile DevicesCode3
BiLLM: Pushing the Limit of Post-Training Quantization for LLMsCode3
Modular Duality in Deep LearningCode3
Allo: A Programming Model for Composable Accelerator DesignCode3
MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of ExpertsCode3
Biomedical and Clinical English Model Packages in the Stanza Python NLP LibraryCode3
mlpack 3: a fast, flexible machine learning libraryCode3
MetaDE: Evolving Differential Evolution by Differential EvolutionCode3
MegaBlocks: Efficient Sparse Training with Mixture-of-ExpertsCode3
Accelerating Goal-Conditioned RL Algorithms and ResearchCode3
M+: Extending MemoryLLM with Scalable Long-Term MemoryCode3
Dataset Distillation with Neural Characteristic Function: A Minmax PerspectiveCode3
Data Generation for Hardware-Friendly Post-Training QuantizationCode3
A GPU-specialized Inference Parameter Server for Large-Scale Deep Recommendation ModelsCode3
MA-LMM: Memory-Augmented Large Multimodal Model for Long-Term Video UnderstandingCode3
CtrLoRA: An Extensible and Efficient Framework for Controllable Image GenerationCode3
Machine Learning in Python: Main developments and technology trends in data science, machine learning, and artificial intelligenceCode3
MagicPIG: LSH Sampling for Efficient LLM GenerationCode3
Tensorized NeuroEvolution of Augmenting Topologies for GPU AccelerationCode3
Merlin: A Vision Language Foundation Model for 3D Computed TomographyCode3
CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning ModelsCode2
A Dataset and Explorer for 3D Signed Distance FunctionsCode2
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMsCode2
LiteTransformerSearch: Training-free Neural Architecture Search for Efficient Language ModelsCode2
AutoTriton: Automatic Triton Programming with Reinforcement Learning in LLMsCode2
LightSeq2: Accelerated Training for Transformer-based Models on GPUsCode2
LightGen: Efficient Image Generation through Knowledge Distillation and Direct Preference OptimizationCode2
LightSeq: A High Performance Inference Library for TransformersCode2
λ-ECLIPSE: Multi-Concept Personalized Text-to-Image Diffusion Models by Leveraging CLIP Latent SpaceCode2
Learning to Walk in Minutes Using Massively Parallel Deep Reinforcement LearningCode2
Learning to Fly in SecondsCode2
DISTFLASHATTN: Distributed Memory-efficient Attention for Long-context LLMs TrainingCode2
LAMP: Learn A Motion Pattern for Few-Shot-Based Video GenerationCode2
Latent Neural Operator for Solving Forward and Inverse PDE ProblemsCode2
3DGen: Triplane Latent Diffusion for Textured Mesh GenerationCode2
AutoFocus: Efficient Multi-Scale InferenceCode2
KAD: No More FAD! An Effective and Efficient Evaluation Metric for Audio GenerationCode2
JAX MD: A Framework for Differentiable PhysicsCode2
JAX, M.D.: A Framework for Differentiable PhysicsCode2
Confucius3-Math: A Lightweight High-Performance Reasoning LLM for Chinese K-12 Mathematics LearningCode2
A User's Guide to KSig: GPU-Accelerated Computation of the Signature KernelCode2
Is One GPU Enough? Pushing Image Generation at Higher-Resolutions with Foundation ModelsCode2
Isaac Gym: High Performance GPU-Based Physics Simulation For Robot LearningCode2
MODNet: Real-Time Trimap-Free Portrait Matting via Objective DecompositionCode2
Instant Volumetric Head AvatarsCode2
INT-FlashAttention: Enabling Flash Attention for INT8 QuantizationCode2
InPars Toolkit: A Unified and Reproducible Synthetic Data Generation Pipeline for Neural Information RetrievalCode2
Invertible Diffusion Models for Compressed SensingCode2
JaxMARL: Multi-Agent RL Environments and Algorithms in JAXCode2
LeanDojo: Theorem Proving with Retrieval-Augmented Language ModelsCode2
Show:102550
← PrevPage 6 of 113Next →

No leaderboard results yet.