SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 851900 of 1356 papers

TitleStatusHype
Data-Free Knowledge Distillation for Image Super-ResolutionCode0
Quantized Neural Networks via -1, +1 Encoding Decomposition and AccelerationCode0
How does topology of neural architectures impact gradient propagation and model performance?Code0
Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and BetterCode1
Topology Distillation for Recommender System0
Masked Training of Neural Networks with Partial Gradients0
A Winning Hand: Compressing Deep Networks Can Improve Out-Of-Distribution RobustnessCode1
Energy-efficient Knowledge Distillation for Spiking Neural Networks0
Heterogeneous Federated Learning using Dynamic Model Pruning and Adaptive Gradient0
ModelDiff: Testing-Based DNN Similarity Comparison for Model Reuse DetectionCode1
FedNILM: Applying Federated Learning to NILM Applications at the Edge0
FedNL: Making Newton-Type Methods Applicable to Federated Learning0
Feature Flow Regularization: Improving Structured Sparsity in Deep Neural Networks0
Bidirectional Distillation for Top-K Recommender SystemCode1
You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature GradientCode1
Enabling Lightweight Fine-tuning for Pre-trained Language Model Compression based on Matrix Product OperatorsCode1
One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers0
Energy-Efficient Model Compression and Splitting for Collaborative Inference Over Time-Varying Channels0
On Attention Redundancy: A Comprehensive Study0
NAS-BERT: Task-Agnostic and Adaptive-Size BERT Compression with Neural Architecture Search0
Towards Efficient Full 8-bit Integer DNN Online Training on Resource-limited Devices without Batch Normalization0
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving GeneralizationCode1
Differentiable Sparsification for Deep Neural Networks0
Model Compression0
How to Explain Neural Networks: an Approximation Perspective0
Clustered Sampling: Low-Variance and Improved Representativity for Clients Selection in Federated LearningCode1
3U-EdgeAI: Ultra-Low Memory Training, Ultra-Low BitwidthQuantization, and Ultra-Low Latency Acceleration0
Test-Time Adaptation Toward Personalized Speech Enhancement: Zero-Shot Learning with Knowledge Distillation0
Neural 3D Scene Compression via Model Compression0
Encoding Weights of Irregular Sparsity for Fixed-to-Fixed Model Compression0
Modulating Regularization Frequency for Efficient Compression-Aware Model Training0
Initialization and Regularization of Factorized Neural LayersCode1
Knowledge Distillation for Swedish NER models: A Search for Performance and Efficiency0
On the Adversarial Robustness of Quantized Neural Networks0
Stealthy Backdoors as Compression ArtifactsCode0
Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer0
Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks0
Skip-Convolutions for Efficient Video ProcessingCode1
Knowledge Distillation as Semiparametric InferenceCode0
Differentiable Model Compression via Pseudo Quantization NoiseCode1
Compact CNN Structure Learning by Knowledge Distillation0
Augmenting Deep Classifiers with Polynomial Neural NetworksCode0
Annealing Knowledge DistillationCode0
Dual Discriminator Adversarial Distillation for Data-free Model Compression0
Reversible Watermarking in Deep Convolutional Neural Networks for Integrity Authentication0
Efficient Personalized Speech Enhancement through Self-Supervised Learning0
Model Compression for Dynamic Forecast CombinationCode0
Tight Compression: Compressing CNN Through Fine-Grained Pruning and Weight Permutation for Efficient Implementation0
Deep Compression for PyTorch Model Deployment on MicrocontrollersCode1
Shrinking Bigfoot: Reducing wav2vec 2.0 footprint0
Show:102550
← PrevPage 18 of 28Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified