SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 951975 of 1356 papers

TitleStatusHype
Modulating Regularization Frequency for Efficient Compression-Aware Model Training0
Encoding Weights of Irregular Sparsity for Fixed-to-Fixed Model Compression0
On the Adversarial Robustness of Quantized Neural Networks0
Knowledge Distillation for Swedish NER models: A Search for Performance and Efficiency0
Stealthy Backdoors as Compression ArtifactsCode0
Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer0
Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks0
Knowledge Distillation as Semiparametric InferenceCode0
Compact CNN Structure Learning by Knowledge Distillation0
Augmenting Deep Classifiers with Polynomial Neural NetworksCode0
Annealing Knowledge DistillationCode0
Dual Discriminator Adversarial Distillation for Data-free Model Compression0
Reversible Watermarking in Deep Convolutional Neural Networks for Integrity Authentication0
Model Compression for Dynamic Forecast CombinationCode0
Efficient Personalized Speech Enhancement through Self-Supervised Learning0
Tight Compression: Compressing CNN Through Fine-Grained Pruning and Weight Permutation for Efficient Implementation0
Shrinking Bigfoot: Reducing wav2vec 2.0 footprint0
Prototype-based Personalized Pruning0
Compacting Deep Neural Networks for Internet of Things: Methods and Applications0
Robust Model Compression Using Deep HypothesesCode0
MWQ: Multiscale Wavelet Quantized Neural Networks0
Formalizing Generalization and Robustness of Neural Networks to Weight Perturbations0
On the Utility of Gradient Compression in Distributed Training SystemsCode0
PURSUhInT: In Search of Informative Hint Points Based on Layer Clustering for Knowledge Distillation0
Preserved central model for faster bidirectional compression in distributed settingsCode0
Show:102550
← PrevPage 39 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified