SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 451475 of 1356 papers

TitleStatusHype
Training dynamic models using early exits for automatic speech recognition on resource-constrained devicesCode0
Pruning Large Language Models via Accuracy Predictor0
Two-Step Knowledge Distillation for Tiny Speech Enhancement0
CoLLD: Contrastive Layer-to-layer Distillation for Compressing Multilingual Pre-trained Speech Encoders0
Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization0
Norm Tweaking: High-performance Low-bit Quantization of Large Language Models0
Compressing Vision Transformers for Low-Resource Visual LearningCode0
ADC/DAC-Free Analog Acceleration of Deep Neural Networks with Frequency Transformation0
Uncovering the Hidden Cost of Model CompressionCode0
Computation-efficient Deep Learning for Computer Vision: A Survey0
Improving Knowledge Distillation for BERT Models: Loss Functions, Mapping Methods, and Weight Tuning0
OmniQuant: Omnidirectionally Calibrated Quantization for Large Language ModelsCode2
DLIP: Distilling Language-Image Pre-training0
QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D Object Detection0
Learning Disentangled Representation with Mutual Information Maximization for Real-Time UAV Tracking0
An Empirical Study of CLIP for Text-based Person SearchCode1
SHARK: A Lightweight Model Compression Approach for Large-scale Recommender Systems0
Diffusion Models for Image Restoration and Enhancement -- A Comprehensive SurveyCode2
Spike-and-slab shrinkage priors for structurally sparse Bayesian neural networks0
Benchmarking Adversarial Robustness of Compressed Deep Learning Models0
Shortcut-V2V: Compression Framework for Video-to-Video Translation based on Temporal Redundancy Reduction0
A Survey on Model Compression for Large Language Models0
FedEdge AI-TC: A Semi-supervised Traffic Classification Method based on Trusted Federated Deep Learning for Mobile Edge Computing0
Resource Constrained Model Compression via Minimax Optimization for Spiking Neural NetworksCode0
Accurate Retraining-free Pruning for Pretrained Encoder-based Language ModelsCode1
Show:102550
← PrevPage 19 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified