SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 11011150 of 1356 papers

TitleStatusHype
Role of Mixup in Topological Persistence Based Knowledge Distillation for Wearable Sensor Data0
Two-Step Knowledge Distillation for Tiny Speech Enhancement0
Runtime Tunable Tsetlin Machines for Edge Inference on eFPGAs0
When Compression Meets Model Compression: Memory-Efficient Double Compression for Large Language Models0
UDC: Unified DNAS for Compressible TinyML Models0
MobiSR: Efficient On-Device Super-Resolution through Heterogeneous Mobile Processors0
SaleNet: A low-power end-to-end CNN accelerator for sustained attention level evaluation using EEG0
Adaptive Quantization of Neural Networks0
Saten: Sparse Augmented Tensor Networks for Post-Training Compression of Large Language Models0
Scalable Teacher Forcing Network for Semi-Supervised Large Scale Data Streams0
Scaling Laws for Deep Learning0
SCSP: Spectral Clustering Filter Pruning with Soft Self-adaption Manners0
SDQ: Sparse Decomposed Quantization for LLM Inference0
Search for Better Students to Learn Distilled Knowledge0
Adaptive Neural Connections for Sparsity Learning0
3U-EdgeAI: Ultra-Low Memory Training, Ultra-Low BitwidthQuantization, and Ultra-Low Latency Acceleration0
SeKron: A Decomposition Method Supporting Many Factorization Structures0
Understanding and Improving Knowledge Distillation0
Selective Convolutional Units: Improving CNNs via Channel Selectivity0
XAI-BayesHAR: A novel Framework for Human Activity Recognition with Integrated Uncertainty and Shapely Values0
Self-calibration for Language Model Quantization and Pruning0
Understanding LLMs: A Comprehensive Overview from Training to Inference0
Self-Supervised Generative Adversarial Compression0
Efficient Personalized Speech Enhancement through Self-Supervised Learning0
YANMTT: Yet Another Neural Machine Translation Toolkit0
Semantic Retention and Extreme Compression in LLMs: Can We Have Both?0
Semantics Prompting Data-Free Quantization for Low-Bit Vision Transformers0
Deep Face Recognition Model Compression via Knowledge Transfer and Distillation0
SEOFP-NET: Compression and Acceleration of Deep Neural Networks for Speech Enhancement Using Sign-Exponent-Only Floating-Points0
Sequence-Level Knowledge Distillation for Model Compression of Attention-based Sequence-to-Sequence Speech Recognition0
Encoding Weights of Irregular Sparsity for Fixed-to-Fixed Model Compression0
Serving and Optimizing Machine Learning Workflows on Heterogeneous Infrastructures0
SGAD: Soft-Guided Adaptively-Dropped Neural Network0
Unleashing Channel Potential: Space-Frequency Selection Convolution for SAR Object Detection0
Unraveling Key Factors of Knowledge Distillation0
SHARK: A Lightweight Model Compression Approach for Large-scale Recommender Systems0
Heterogeneous Federated Learning using Dynamic Model Pruning and Adaptive Gradient0
Shortcut-V2V: Compression Framework for Video-to-Video Translation based on Temporal Redundancy Reduction0
Adapting Models to Signal Degradation using Distillation0
Shrinking Bigfoot: Reducing wav2vec 2.0 footprint0
ShrinkML: End-to-End ASR Model Compression Using Reinforcement Learning0
Unsupervised model compression for multilayer bootstrap networks0
Simplifying Two-Stage Detectors for On-Device Inference in Remote Sensing0
Masked Training of Neural Networks with Partial Gradients0
UPAQ: A Framework for Real-Time and Energy-Efficient 3D Object Detection in Autonomous Vehicles0
AdapMTL: Adaptive Pruning Framework for Multitask Learning Model0
3DQ: Compact Quantized Neural Networks for Volumetric Whole Brain Segmentation0
USDC: Unified Static and Dynamic Compression for Visual Transformer0
AdaKD: Dynamic Knowledge Distillation of ASR models using Adaptive Loss Weighting0
Small, Accurate, and Fast Vehicle Re-ID on the Edge: the SAFR Approach0
Show:102550
← PrevPage 23 of 28Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified