SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 11711180 of 1356 papers

TitleStatusHype
Towards Understanding and Improving Knowledge Distillation for Neural Machine TranslationCode0
Occam Gradient DescentCode0
CoDiNet: Path Distribution Modeling with Consistency and Diversity for Dynamic RoutingCode0
HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model CompressionCode0
Is Smaller Always Faster? Tradeoffs in Compressing Self-Supervised Speech TransformersCode0
RemoteTrimmer: Adaptive Structural Pruning for Remote Sensing Image ClassificationCode0
A Programmable Approach to Neural Network CompressionCode0
Beyond Perplexity: Multi-dimensional Safety Evaluation of LLM CompressionCode0
How does topology of neural architectures impact gradient propagation and model performance?Code0
Compressing Convolutional Neural Networks via Factorized Convolutional FiltersCode0
Show:102550
← PrevPage 118 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified