SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 931940 of 1356 papers

TitleStatusHype
Fast convolutional neural networks on FPGAs with hls4mlCode2
Activation Density based Mixed-Precision Quantization for Energy Efficient Neural Networks0
Adversarially Robust and Explainable Model Compression with On-Device Personalization for Text Classification0
On-Device Document Classification using multimodal features0
Improving Neural Network Efficiency via Post-Training Quantization With Adaptive Floating-PointCode1
Exploration and Estimation for Model Compression0
Improve Object Detection with Feature-based Knowledge Distillation: Towards Accurate and Efficient DetectorsCode1
Model Compression via Hyper-Structure Network0
Knowledge distillation via softmax regression representation learning0
Post-Training Weighted Quantization of Neural Networks for Language Models0
Show:102550
← PrevPage 94 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified