SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 871880 of 1356 papers

TitleStatusHype
Robustness Challenges in Model Distillation and Pruning for Natural Language Understanding0
What do larger image classifiers memorise?0
What is Left After Distillation? How Knowledge Transfer Impacts Fairness and Bias0
What is Lost in Knowledge Distillation?0
What Makes a Good Dataset for Knowledge Distillation?0
When Compression Meets Model Compression: Memory-Efficient Double Compression for Large Language Models0
XAI-BayesHAR: A novel Framework for Human Activity Recognition with Integrated Uncertainty and Shapely Values0
YANMTT: Yet Another Neural Machine Translation Toolkit0
You Only Prune Once: Designing Calibration-Free Model Compression With Policy Learning0
Individual Content and Motion Dynamics Preserved Pruning for Video Diffusion Models0
Show:102550
← PrevPage 88 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified