SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 426450 of 1356 papers

TitleStatusHype
Empirical Evaluation of Deep Learning Model Compression Techniques on the WaveNet VocoderCode0
ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer AccelerationCode0
An exploration of the effect of quantisation on energy consumption and inference time of StarCoder2Code0
Causal Explanation of Convolutional Neural NetworksCode0
StrassenNets: Deep Learning with a Multiplication BudgetCode0
Einconv: Exploring Unexplored Tensor Network Decompositions for Convolutional Neural NetworksCode0
Improved Knowledge Distillation via Full Kernel Matrix TransferCode0
Compressing Convolutional Neural Networks via Factorized Convolutional FiltersCode0
Efficient model compression with Random Operation Access Specific Tile (ROAST) hashingCode0
Efficient Speech Translation through Model Compression and Knowledge DistillationCode0
Systematic Outliers in Large Language ModelsCode0
Exploring Unexplored Tensor Network Decompositions for Convolutional Neural NetworksCode0
Tensorized Embedding Layers for Efficient Model CompressionCode0
Learning Accurate Performance Predictors for Ultrafast Automated Model CompressionCode0
On Model Compression for Neural Networks: Framework, Algorithm, and Convergence GuaranteeCode0
Compressed models are NOT miniature versions of large models0
Artemis: HE-Aware Training for Efficient Privacy-Preserving Machine Learning0
Comprehensive Survey of Model Compression and Speed up for Vision Transformers0
Are We There Yet? A Measurement Study of Efficiency for LLM Applications on Mobile Devices0
Comprehensive Study on Performance Evaluation and Optimization of Model Compression: Bridging Traditional Deep Learning and Large Language Models0
Compositionality Unlocks Deep Interpretable Models0
A Comprehensive Review and a Taxonomy of Edge Machine Learning: Requirements, Paradigms, and Techniques0
Accelerating Very Deep Convolutional Networks for Classification and Detection0
CompMarkGS: Robust Watermarking for Compressed 3D Gaussian Splatting0
Complexity-Driven CNN Compression for Resource-constrained Edge AI0
Show:102550
← PrevPage 18 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified