SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 711720 of 1356 papers

TitleStatusHype
An Empirical Study of Low Precision Quantization for TinyML0
Don't Be So Dense: Sparse-to-Sparse GAN Training Without Sacrificing Performance0
Structured Pruning is All You Need for Pruning CNNs at Initialization0
E-LANG: Energy-Based Joint Inferencing of Super and Swift Language Models0
KMIR: A Benchmark for Evaluating Knowledge Memorization, Identification and Reasoning Abilities of Language Models0
Multi-task Learning Approach for Modulation and Wireless Signal Classification for 5G and Beyond: Edge Deployment via Model Compression0
A Novel Architecture Slimming Method for Network Pruning and Knowledge Distillation0
Time-Correlated Sparsification for Efficient Over-the-Air Model Aggregation in Wireless Federated Learning0
A Survey on Model Compression and Acceleration for Pretrained Language Models0
SPDY: Accurate Pruning with Speedup GuaranteesCode1
Show:102550
← PrevPage 72 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified