SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 171180 of 1356 papers

TitleStatusHype
Passport-aware Normalization for Deep Model ProtectionCode1
CompRess: Self-Supervised Learning by Compressing RepresentationsCode1
Towards Compact Neural Networks via End-to-End Training: A Bayesian Tensor Approach with Automatic Rank DeterminationCode1
BERT-EMD: Many-to-Many Layer Mapping for BERT Compression with Earth Mover's DistanceCode1
Contrastive Distillation on Intermediate Representations for Language Model CompressionCode1
Densely Guided Knowledge Distillation using Multiple Teacher AssistantsCode1
Implicit Regularization via Neural Feature AlignmentCode1
Paying more attention to snapshots of Iterative Pruning: Improving Model Compression via Ensemble DistillationCode1
Improving Post Training Neural Quantization: Layer-wise Calibration and Integer ProgrammingCode1
Knowledge Distillation Meets Self-SupervisionCode1
Show:102550
← PrevPage 18 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified