SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 861870 of 1356 papers

TitleStatusHype
Vision Transformers on the Edge: A Comprehensive Survey of Model Compression and Acceleration Strategies0
VQ4ALL: Efficient Neural Network Representation via a Universal Codebook0
Wasserstein Contrastive Representation Distillation0
Watermarking Graph Neural Networks by Random Graphs0
WeClick: Weakly-Supervised Video Semantic Segmentation with Click Annotations0
Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators0
Weight Normalization based Quantization for Deep Neural Network Compression0
Weight Squeezing: Reparameterization for Knowledge Transfer and Model Compression0
Weight Squeezing: Reparameterization for Compression and Fast Inference0
Weight Squeezing: Reparameterization for Knowledge Transfer and Model Compression0
Show:102550
← PrevPage 87 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified