SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 110 of 1356 papers

TitleStatusHype
LINR-PCGC: Lossless Implicit Neural Representations for Point Cloud Geometry Compression0
DipSVD: Dual-importance Protected SVD for Efficient LLM Compression0
RLRC: Reinforcement Learning-based Recovery for Compressed Vision-Language-Action Models0
Model compression using knowledge distillation with integrated gradients0
Attribution-guided Pruning for Compression, Circuit Discovery, and Targeted Correction in LLMsCode0
Simple is what you need for efficient and accurate medical image segmentationCode0
EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware OptimizationCode0
FIMA-Q: Post-Training Quantization for Vision Transformers by Fisher Information Matrix ApproximationCode1
Advances in Small-Footprint Keyword Spotting: A Comprehensive Review of Efficient Models and AlgorithmsCode0
Post-Training Quantization for Video Matting0
Show:102550
← PrevPage 1 of 136Next →

Benchmark Results