SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 821830 of 1356 papers

TitleStatusHype
Towards Better Parameter-Efficient Fine-Tuning for Large Language Models: A Position Paper0
Towards Building a Real Time Mobile Device Bird Counting System Through Synthetic Data Training and Model Compression0
Towards domain generalisation in ASR with elitist sampling and ensemble knowledge distillation0
Towards efficient deep autoencoders for multivariate time series anomaly detection0
Towards Efficient Deep Spiking Neural Networks Construction with Spiking Activity based Pruning0
Towards Efficient Full 8-bit Integer DNN Online Training on Resource-limited Devices without Batch Normalization0
Towards Efficient Tensor Decomposition-Based DNN Model Compression with Optimization Framework0
Towards Higher Ranks via Adversarial Weight Pruning0
Towards Modality Transferable Visual Information Representation with Optimal Model Compression0
Towards Optimal Compression: Joint Pruning and Quantization0
Show:102550
← PrevPage 83 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified