SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 12611270 of 1356 papers

TitleStatusHype
Chemical transformer compression for accelerating both training and inference of molecular modelingCode0
A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNsCode0
Training dynamic models using early exits for automatic speech recognition on resource-constrained devicesCode0
PENNI: Pruned Kernel Sharing for Efficient CNN InferenceCode0
subMFL: Compatiple subModel Generation for Federated Learning in Device Heterogenous EnvironmentCode0
Perforated Backpropagation: A Neuroscience Inspired Extension to Artificial Neural NetworksCode0
Exploring Gradient Flow Based Saliency for DNN Model CompressionCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
LXMERT Model Compression for Visual Question AnsweringCode0
Deep Neural Network Compression for Image Classification and Object DetectionCode0
Show:102550
← PrevPage 127 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified