SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 676700 of 1356 papers

TitleStatusHype
A Contrastive Knowledge Transfer Framework for Model Compression and Transfer LearningCode0
OTOV2: Automatic, Generic, User-Friendly0
On Model Compression for Neural Networks: Framework, Algorithm, and Convergence GuaranteeCode0
Greener yet Powerful: Taming Large Code Generation Models with Quantization0
Gradient-Free Structured Pruning with Unlabeled Data0
Rotation Invariant Quantization for Model CompressionCode0
Adversarial Attacks on Machine Learning in Embedded and IoT Platforms0
Towards domain generalisation in ASR with elitist sampling and ensemble knowledge distillation0
Debiased Distillation by Transplanting the Last Layer0
Structured Bayesian Compression for Deep Neural Networks Based on The Turbo-VBI Approach0
HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained Transformers0
A Comprehensive Review and a Taxonomy of Edge Machine Learning: Requirements, Paradigms, and Techniques0
Towards Optimal Compression: Joint Pruning and Quantization0
On Achieving Privacy-Preserving State-of-the-Art Edge Intelligence0
Knowledge Distillation in Vision Transformers: A Critical Review0
Generalized Uncertainty of Deep Neural Networks: Taxonomy and Applications0
Knowledge Distillation on Graphs: A Survey0
AMD: Adaptive Masked Distillation for Object Detection0
Improved knowledge distillation by utilizing backward pass knowledge in neural networks0
HALOC: Hardware-Aware Automatic Low-Rank Compression for Compact Neural Networks0
Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI FeedbackCode0
HCE: Improving Performance and Efficiency with Heterogeneously Compressed Neural Network Ensemble0
Distilling Focal Knowledge From Imperfect Expert for 3D Object DetectionCode0
One-Shot Model for Mixed-Precision Quantization0
Tiny Updater: Towards Efficient Neural Network-Driven Software UpdatingCode0
Show:102550
← PrevPage 28 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified