SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 771780 of 1356 papers

TitleStatusHype
Oracle Teacher: Leveraging Target Information for Better Knowledge Distillation of CTC Models0
Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators0
How to Select One Among All ? An Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding0
Distilling Object Detectors with Feature RichnessCode1
ILMPQ : An Intra-Layer Multi-Precision Deep Neural Network Quantization framework for FPGA0
On Cross-Layer Alignment for Model Fusion of Heterogeneous Neural Networks0
Generalized Depthwise-Separable Convolutions for Adversarially Robust and Efficient Neural NetworksCode1
Reconstructing Pruned Filters using Cheap Spatial Transformations0
Exploring Gradient Flow Based Saliency for DNN Model CompressionCode0
How and When Adversarial Robustness Transfers in Knowledge Distillation?0
Show:102550
← PrevPage 78 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified