SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 731740 of 1356 papers

TitleStatusHype
Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization0
Learning Disentangled Representation with Mutual Information Maximization for Real-Time UAV Tracking0
Three Dimensional Convolutional Neural Network Pruning with Regularization-Based Method0
Learning Efficient Image Super-Resolution Networks via Structure-Regularized Pruning0
Learning Efficient Object Detection Models with Knowledge Distillation0
ASCAI: Adaptive Sampling for acquiring Compact AI0
Learning by Sampling and Compressing: Efficient Graph Representation Learning with Extremely Limited Annotations0
MixMix: All You Need for Data-Free Compression Are Feature and Data Mixing0
Learning Interpretation with Explainable Knowledge Distillation0
WeClick: Weakly-Supervised Video Semantic Segmentation with Click Annotations0
Show:102550
← PrevPage 74 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified