SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 11011125 of 1356 papers

TitleStatusHype
Role of Mixup in Topological Persistence Based Knowledge Distillation for Wearable Sensor Data0
Two-Step Knowledge Distillation for Tiny Speech Enhancement0
Runtime Tunable Tsetlin Machines for Edge Inference on eFPGAs0
When Compression Meets Model Compression: Memory-Efficient Double Compression for Large Language Models0
UDC: Unified DNAS for Compressible TinyML Models0
MobiSR: Efficient On-Device Super-Resolution through Heterogeneous Mobile Processors0
SaleNet: A low-power end-to-end CNN accelerator for sustained attention level evaluation using EEG0
Adaptive Quantization of Neural Networks0
Saten: Sparse Augmented Tensor Networks for Post-Training Compression of Large Language Models0
Scalable Teacher Forcing Network for Semi-Supervised Large Scale Data Streams0
Scaling Laws for Deep Learning0
SCSP: Spectral Clustering Filter Pruning with Soft Self-adaption Manners0
SDQ: Sparse Decomposed Quantization for LLM Inference0
Search for Better Students to Learn Distilled Knowledge0
Adaptive Neural Connections for Sparsity Learning0
3U-EdgeAI: Ultra-Low Memory Training, Ultra-Low BitwidthQuantization, and Ultra-Low Latency Acceleration0
SeKron: A Decomposition Method Supporting Many Factorization Structures0
Understanding and Improving Knowledge Distillation0
Selective Convolutional Units: Improving CNNs via Channel Selectivity0
XAI-BayesHAR: A novel Framework for Human Activity Recognition with Integrated Uncertainty and Shapely Values0
Self-calibration for Language Model Quantization and Pruning0
Understanding LLMs: A Comprehensive Overview from Training to Inference0
Self-Supervised Generative Adversarial Compression0
Efficient Personalized Speech Enhancement through Self-Supervised Learning0
YANMTT: Yet Another Neural Machine Translation Toolkit0
Show:102550
← PrevPage 45 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified