SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 13011325 of 1356 papers

TitleStatusHype
Efficient Recurrent Neural Networks using Structured Matrices in FPGAs0
Interpreting Deep Classifier by Visual Distillation of Dark Knowledge0
Model compression via distillation and quantizationCode0
Paraphrasing Complex Network: Network Compression via Factor TransferCode0
AMC: AutoML for Model Compression and Acceleration on Mobile DevicesCode2
Model compression for faster structural separation of macromolecules captured by Cellular Electron Cryo-Tomography0
DNN Model Compression Under Accuracy Constraints0
Don't encrypt the data; just approximate the model \ Towards Secure Transaction and Fair Pricing of Training Data0
Adaptive Quantization of Neural Networks0
Learning Deep and Compact Models for Gesture RecognitionCode0
StrassenNets: Deep Learning with a Multiplication BudgetCode0
Learning Efficient Object Detection Models with Knowledge Distillation0
Neural Regularized Domain Adaptation for Chinese Word Segmentation0
MicroExpNet: An Extremely Small and Fast Model For Expression Recognition From Face ImagesCode0
Improved Bayesian Compression0
Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy0
Weightless: Lossy Weight Encoding For Deep Neural Network CompressionCode0
A Survey of Model Compression and Acceleration for Deep Neural Networks0
Data-Free Knowledge Distillation for Deep Neural NetworksCode2
To prune, or not to prune: exploring the efficacy of pruning for model compressionCode1
Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks0
N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning0
Learning Intrinsic Sparse Structures within Long Short-Term MemoryCode0
A Deep Cascade Network for Unaligned Face Attribute Classification0
Model Distillation with Knowledge Transfer from Face Classification to Alignment and Verification0
Show:102550
← PrevPage 53 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified