SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 601625 of 1356 papers

TitleStatusHype
Greener yet Powerful: Taming Large Code Generation Models with Quantization0
Group channel pruning and spatial attention distilling for object detection0
GroupReduce: Block-Wise Low-Rank Approximation for Neural Language Model Shrinking0
Atrial Fibrillation Detection Using Weight-Pruned, Log-Quantised Convolutional Neural Networks0
Fragile Mastery: Are Domain-Specific Trade-Offs Undermining On-Device Language Models?0
HadaNets: Flexible Quantization Strategies for Neural Networks0
HALOC: Hardware-Aware Automatic Low-Rank Compression for Compact Neural Networks0
Conditional Automated Channel Pruning for Deep Neural Networks0
HCE: Improving Performance and Efficiency with Heterogeneously Compressed Neural Network Ensemble0
A flexible, extensible software framework for model compression based on the LC algorithm0
HFSP: A Hardware-friendly Soft Pruning Framework for Vision Transformers0
HideNseek: Federated Lottery Ticket via Server-side Pruning and Sign Supermask0
Cross-Channel Intragroup Sparsity Neural Network0
Attention Sinks and Outlier Features: A 'Catch, Tag, and Release' Mechanism for Embeddings0
Improving Acoustic Scene Classification in Low-Resource Conditions0
HODEC: Towards Efficient High-Order DEcomposed Convolutional Neural Networks0
ConaCLIP: Exploring Distillation of Fully-Connected Knowledge Interaction Graph for Lightweight Text-Image Retrieval0
How and When Adversarial Robustness Transfers in Knowledge Distillation?0
On the Impact of Calibration Data in Post-training Quantization and Pruning0
Deep Face Recognition Model Compression via Knowledge Transfer and Distillation0
How to Explain Neural Networks: an Approximation Perspective0
How to Select One Among All ? An Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding0
Formalizing Generalization and Robustness of Neural Networks to Weight Perturbations0
Aerial Image Classification in Scarce and Unconstrained Environments via Conformal Prediction0
Formalizing Generalization and Adversarial Robustness of Neural Networks to Weight Perturbations0
Show:102550
← PrevPage 25 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified