SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 381390 of 1356 papers

TitleStatusHype
Large receptive field strategy and important feature extraction strategy in 3D object detection0
Pruning for Protection: Increasing Jailbreak Resistance in Aligned LLMs Without Fine-TuningCode0
ELRT: Efficient Low-Rank Training for Compact Convolutional Neural Networks0
Model Compression Techniques in Biometrics Applications: A SurveyCode0
SymbolNet: Neural Symbolic Regression with Adaptive Dynamic Pruning for CompressionCode0
Dynamic DNNs and Runtime Management for Efficient Inference on Mobile/Embedded DevicesCode1
Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning0
Knowledge Translation: A New Pathway for Model CompressionCode0
FFSplit: Split Feed-Forward Network For Optimizing Accuracy-Efficiency Trade-off in Language Model Inference0
Understanding LLMs: A Comprehensive Overview from Training to Inference0
Show:102550
← PrevPage 39 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified