SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 13261350 of 1356 papers

TitleStatusHype
DeepRebirth: Accelerating Deep Neural Network Execution on Mobile Devices0
Model compression as constrained optimization, with application to neural nets. Part II: quantization0
Model compression as constrained optimization, with application to neural nets. Part I: general framework0
DarkRank: Accelerating Deep Metric Learning via Cross Sample Similarities Transfer0
Tensor Contraction Layers for Parsimonious Deep Nets0
Cross-lingual Distillation for Text ClassificationCode0
Acoustic Model Compression with MAP adaptation0
Exploiting random projections and sparsity with random forests and gradient boosting methods -- Application to multi-label and multi-output learning, random forest model compression and leveraging input sparsity0
A Compact DNN: Approaching GoogLeNet-Level Accuracy of Classification and Domain Adaptation0
Exploiting Domain Knowledge via Grouped Weight Sharing with Application to Text Categorization0
Compression of Deep Neural Networks for Image Instance Retrieval0
QuickNet: Maximizing Efficiency and Efficacy in Deep Architectures0
Two-Bit Networks for Deep Learning on Resource-Constrained Embedded Devices0
Parameter Compression of Recurrent Neural Networks and Degradation of Short-term Memory0
The Shallow End: Empowering Shallower Deep-Convolutional Networks through Auxiliary OutputsCode0
Deep Model Compression: Distilling Knowledge from Noisy Teachers0
Ensemble-Compression: A New Method for Parallel Training of Deep Neural Networks0
Ternary Weight NetworksCode1
Adapting Models to Signal Degradation using Distillation0
On the Compression of Recurrent Neural Networks with an Application to LVCSR acoustic modeling for Embedded Speech Recognition0
SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model sizeCode1
Actor-Mimic: Deep Multitask and Transfer Reinforcement LearningCode0
Blending LSTMs into CNNs0
Distilling Model KnowledgeCode0
DeepFont: Identify Your Font from An ImageCode0
Show:102550
← PrevPage 54 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified