SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 421430 of 1356 papers

TitleStatusHype
CoDiNet: Path Distribution Modeling with Consistency and Diversity for Dynamic RoutingCode0
Adversarial Fine-tuning of Compressed Neural Networks for Joint Improvement of Robustness and EfficiencyCode0
Enhancing Knowledge Distillation of Large Language Models through Efficient Multi-Modal Distribution AlignmentCode0
Explicit-NeRF-QA: A Quality Assessment Database for Explicit NeRF Model CompressionCode0
Einconv: Exploring Unexplored Tensor Network Decompositions for Convolutional Neural NetworksCode0
ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer AccelerationCode0
CASP: Compression of Large Multimodal Models Based on Attention SparsityCode0
An exploration of the effect of quantisation on energy consumption and inference time of StarCoder2Code0
Causal Explanation of Convolutional Neural NetworksCode0
Efficient Speech Translation through Model Compression and Knowledge DistillationCode0
Show:102550
← PrevPage 43 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified