SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 641650 of 1356 papers

TitleStatusHype
Improve Knowledge Distillation via Label Revision and Data Selection0
Interpreting Deep Classifier by Visual Distillation of Dark Knowledge0
Intrinsically Sparse Long Short-Term Memory Networks0
Improving Knowledge Distillation for BERT Models: Loss Functions, Mapping Methods, and Weight Tuning0
Is Quantum Optimization Ready? An Effort Towards Neural Network Compression using Adiabatic Quantum Computing0
FlatENN: Train Flat for Enhanced Fault Tolerance of Quantized Deep Neural Networks0
FIT: A Metric for Model Sensitivity0
In defense of parameter sharing for model-compression0
Individual Content and Motion Dynamics Preserved Pruning for Video Diffusion Models0
Compress then Serve: Serving Thousands of LoRA Adapters with Little Overhead0
Show:102550
← PrevPage 65 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified