SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 11011110 of 1356 papers

TitleStatusHype
Role of Mixup in Topological Persistence Based Knowledge Distillation for Wearable Sensor Data0
Two-Step Knowledge Distillation for Tiny Speech Enhancement0
Runtime Tunable Tsetlin Machines for Edge Inference on eFPGAs0
When Compression Meets Model Compression: Memory-Efficient Double Compression for Large Language Models0
UDC: Unified DNAS for Compressible TinyML Models0
MobiSR: Efficient On-Device Super-Resolution through Heterogeneous Mobile Processors0
SaleNet: A low-power end-to-end CNN accelerator for sustained attention level evaluation using EEG0
Adaptive Quantization of Neural Networks0
Saten: Sparse Augmented Tensor Networks for Post-Training Compression of Large Language Models0
Scalable Teacher Forcing Network for Semi-Supervised Large Scale Data Streams0
Show:102550
← PrevPage 111 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified