SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 411420 of 1356 papers

TitleStatusHype
Towards Higher Ranks via Adversarial Weight Pruning0
Relationship between Model Compression and Adversarial Robustness: A Review of Current Evidence0
Cosine Similarity Knowledge Distillation for Individual Class Information Transfer0
Knowledge Distillation Based Semantic Communications For Multiple Users0
Education distillation:getting student models to learn in shcools0
Efficient Transformer Knowledge Distillation: A Performance Review0
Towards Better Parameter-Efficient Fine-Tuning for Large Language Models: A Position Paper0
Compact 3D Gaussian Representation for Radiance FieldCode2
Shedding the Bits: Pushing the Boundaries of Quantization with Minifloats on FPGAs0
LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient Language Model FinetuningCode1
Show:102550
← PrevPage 42 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified