SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 9511000 of 1356 papers

TitleStatusHype
Modulating Regularization Frequency for Efficient Compression-Aware Model Training0
Encoding Weights of Irregular Sparsity for Fixed-to-Fixed Model Compression0
On the Adversarial Robustness of Quantized Neural Networks0
Knowledge Distillation for Swedish NER models: A Search for Performance and Efficiency0
Stealthy Backdoors as Compression ArtifactsCode0
Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer0
Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks0
Knowledge Distillation as Semiparametric InferenceCode0
Compact CNN Structure Learning by Knowledge Distillation0
Augmenting Deep Classifiers with Polynomial Neural NetworksCode0
Annealing Knowledge DistillationCode0
Dual Discriminator Adversarial Distillation for Data-free Model Compression0
Reversible Watermarking in Deep Convolutional Neural Networks for Integrity Authentication0
Model Compression for Dynamic Forecast CombinationCode0
Efficient Personalized Speech Enhancement through Self-Supervised Learning0
Tight Compression: Compressing CNN Through Fine-Grained Pruning and Weight Permutation for Efficient Implementation0
Shrinking Bigfoot: Reducing wav2vec 2.0 footprint0
Prototype-based Personalized Pruning0
Compacting Deep Neural Networks for Internet of Things: Methods and Applications0
Robust Model Compression Using Deep HypothesesCode0
MWQ: Multiscale Wavelet Quantized Neural Networks0
Formalizing Generalization and Robustness of Neural Networks to Weight Perturbations0
On the Utility of Gradient Compression in Distributed Training SystemsCode0
PURSUhInT: In Search of Informative Hint Points Based on Layer Clustering for Knowledge Distillation0
Preserved central model for faster bidirectional compression in distributed settingsCode0
Lottery Ticket Preserves Weight Correlation: Is It Desirable or Not?0
Neural Network Compression for Noisy Storage Devices0
Robustness in Compressed Neural Networks for Object Detection0
Compressed Object DetectionCode0
It's always personal: Using Early Exits for Efficient On-Device CNN Personalisation0
AACP: Model Compression by Accurate and Automatic Channel Pruning0
Deep Model Compression based on the Training History0
AdaSpring: Context-adaptive and Runtime-evolutionary Deep Model Compression for Mobile Applications0
Differential Privacy Meets Federated Learning under Communication Constraints0
Collaborative Teacher-Student Learning via Multiple Knowledge Transfer0
Deep Compression of Neural Networks for Fault Detection on Tennessee Eastman Chemical Processes0
Model Compression for Domain Adaptation through Causal Effect EstimationCode0
Single-path Bit Sharing for Automatic Loss-aware Model Compression0
Activation Density based Mixed-Precision Quantization for Energy Efficient Neural Networks0
Adversarially Robust and Explainable Model Compression with On-Device Personalization for Text Classification0
On-Device Document Classification using multimodal features0
Block Skim Transformer for Efficient Question Answering0
TwinDNN: A Tale of Two Deep Neural Networks0
Model Compression via Hyper-Structure Network0
Can Students Outperform Teachers in Knowledge Distillation based Model Compression?0
Knowledge distillation via softmax regression representation learning0
SACoD: Sensor Algorithm Co-Design Towards Efficient CNN-powered Intelligent PhlatCamCode0
Dynamic Probabilistic Pruning: Training sparse networks based on stochastic and dynamic masking0
Exploration and Estimation for Model Compression0
Post-Training Weighted Quantization of Neural Networks for Language Models0
Show:102550
← PrevPage 20 of 28Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified