SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 10811090 of 1356 papers

TitleStatusHype
Exploring the Limits of Simple Learners in Knowledge Distillation for Document Classification with DocBERT0
On the Demystification of Knowledge Distillation: A Residual Network Perspective0
PFGDF: Pruning Filter via Gaussian Distribution Feature for Deep Neural Networks Acceleration0
Additive Tree-Structured Covariance Function for Conditional Parameter Spaces in Bayesian Optimization0
ResOT: Resource-Efficient Oblique Trees for Neural Signal Classification0
Dynamic Model Pruning with Feedback0
An Embedded Deep Learning Object Detection Model For Traffic In Asian Countries0
Knowledge Distillation: A Survey0
EDCompress: Energy-Aware Model Compression for Dataflows0
AdaDeep: A Usage-Driven, Automated Deep Model Compression Framework for Enabling Ubiquitous Intelligent Mobiles0
Show:102550
← PrevPage 109 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified