SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 321330 of 1356 papers

TitleStatusHype
Bayesian Federated Model Compression for Communication and Computation Efficiency0
Multilingual Brain Surgeon: Large Language Models Can be Compressed Leaving No Language BehindCode0
Improve Knowledge Distillation via Label Revision and Data Selection0
Knowledge Distillation with Multi-granularity Mixture of Priors for Image Super-Resolution0
On Linearizing Structured Data in Encoder-Decoder Language Models: Insights from Text-to-SQL0
Automated Inference of Graph Transformation Rules0
Enhancing Inference Efficiency of Large Language Models: Investigating Optimization Strategies and Architectural Innovations0
Instance-Aware Group Quantization for Vision Transformers0
Streamlining Redundant Layers to Compress Large Language ModelsCode1
Is Modularity Transferable? A Case Study through the Lens of Knowledge DistillationCode0
Show:102550
← PrevPage 33 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified