SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 11511160 of 1356 papers

TitleStatusHype
Model Compression with Two-stage Multi-teacher Knowledge Distillation for Web Question Answering System0
Compacting, Picking and Growing for Unforgetting Continual LearningCode1
Structured Pruning of a BERT-based Question Answering Model0
Model Fusion via Optimal TransportCode0
Structured Pruning of Large Language ModelsCode1
Differentiable Sparsification for Deep Neural Networks0
Deep Neural Network Compression for Image Classification and Object DetectionCode0
How does topology influence gradient propagation and model performance of deep networks with DenseNet-type skip connections?Code0
Distilled Split Deep Neural Networks for Edge-Assisted Real-Time SystemsCode1
Adversarial Robustness vs. Model Compression, or Both?Code0
Show:102550
← PrevPage 116 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified