SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 10111020 of 1356 papers

TitleStatusHype
Inferring ECG from PPG for Continuous Cardiac Monitoring Using Lightweight Neural Network0
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework0
Model Compression Using Optimal Transport0
Multi-head Knowledge Distillation for Model Compression0
Meta-KD: A Meta Knowledge Distillation Framework for Language Model Compression across Domains0
Compressing Pre-trained Language Models by Matrix Decomposition0
Self-Supervised Generative Adversarial Compression0
NPAS: A Compiler-aware Framework of Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration0
Edge Deep Learning for Neural Implants0
Reverse-engineering recurrent neural network solutions to a hierarchical inference task for mice0
Show:102550
← PrevPage 102 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified