SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 671680 of 1356 papers

TitleStatusHype
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models0
Do we need Label Regularization to Fine-tune Pre-trained Language Models?0
PointDistiller: Structured Knowledge Distillation Towards Efficient and Compact 3D DetectionCode1
Aligning Logits Generatively for Principled Black-Box Knowledge DistillationCode0
InDistill: Information flow-preserving knowledge distillation for model compressionCode0
QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators0
Dimensionality Reduced Training by Pruning and Freezing Parts of a Deep Neural Network, a Survey0
Perturbation of Deep Autoencoder Weights for Model Compression and Classification of Tabular Data0
Chemical transformer compression for accelerating both training and inference of molecular modelingCode0
DNA data storage, sequencing data-carrying DNA0
Show:102550
← PrevPage 68 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified