SOTAVerified

GPU

Papers

Showing 901925 of 5629 papers

TitleStatusHype
Reinforcement learning with learned gadgets to tackle hard quantum problems on real hardwareCode0
A Novel Breast Ultrasound Image Augmentation Method Using Advanced Neural Style Transfer: An Efficient and Explainable Approach0
The Importance of Being Scalable: Improving the Speed and Accuracy of Neural Network Interatomic Potentials Across Chemical DomainsCode2
Cycle-Constrained Adversarial Denoising Convolutional Network for PET Image Denoising: Multi-Dimensional Validation on Large Datasets with Reader Study and Real Low-Dose Data0
Context-Aware Token Selection and Packing for Enhanced Vision Transformer0
Very fast Bayesian Additive Regression Trees on GPUCode2
$100K or 100 Days: Trade-offs when Pre-Training with Academic ResourcesCode2
A Message Passing Neural Network Surrogate Model for Bond-Associated Peridynamic Material Correspondence Formulation0
AI-assisted Agile Propagation Modeling for Real-time Digital Twin Wireless Networks0
Application of Audio Fingerprinting Techniques for Real-Time Scalable Speech Retrieval and Speech Clusterization0
VL-Cache: Sparsity and Modality-Aware KV Cache Compression for Vision-Language Model Inference Acceleration0
Memory-Efficient Point Cloud Registration via Overlapping Region Sampling0
Revisiting Reliability in Large-Scale Machine Learning Research Clusters0
Pushing the Performance Envelope of DNN-based Recommendation Systems Inference on GPUsCode0
Motion Graph Unleashed: A Novel Approach to Video PredictionCode0
Data Generation for Hardware-Friendly Post-Training QuantizationCode3
ProMoE: Fast MoE-based LLM Serving using Proactive Caching0
ShadowKV: KV Cache in Shadows for High-Throughput Long-Context LLM InferenceCode3
FusedInf: Efficient Swapping of DNN Models for On-Demand Serverless Inference Services on the EdgeCode0
Accelerated Bayesian parameter estimation and model selection for gravitational waves with normalizing flows0
Modular Duality in Deep LearningCode3
KD-LoRA: A Hybrid Approach to Efficient Fine-Tuning with LoRA and Knowledge DistillationCode1
ThunderKittens: Simple, Fast, and Adorable AI KernelsCode7
Deep Optimizer States: Towards Scalable Training of Transformer Models Using Interleaved OffloadingCode0
Computational Bottlenecks of Training Small-scale Large Language Models0
Show:102550
← PrevPage 37 of 226Next →

No leaderboard results yet.