SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 5175 of 1356 papers

TitleStatusHype
Compression-Aware Video Super-ResolutionCode1
Differentiable Model Compression via Pseudo Quantization NoiseCode1
Discovering Dynamic Patterns from Spatiotemporal Data with Time-Varying Low-Rank AutoregressionCode1
Discrimination-aware Channel Pruning for Deep Neural NetworksCode1
Distilling Linguistic Context for Language Model CompressionCode1
Distilling Object Detectors with Feature RichnessCode1
DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and TransformersCode1
Dual Relation Knowledge Distillation for Object DetectionCode1
3DG-STFM: 3D Geometric Guided Student-Teacher Feature MatchingCode1
AD-KD: Attribution-Driven Knowledge Distillation for Language Model CompressionCode1
ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Method of MultipliersCode1
EarlyBERT: Efficient BERT Training via Early-bird Lottery TicketsCode1
CompRess: Self-Supervised Learning by Compressing RepresentationsCode1
Designing Large Foundation Models for Efficient Training and Inference: A SurveyCode1
Activation-Informed Merging of Large Language ModelsCode1
Comprehensive Knowledge Distillation with Causal InterventionCode1
Streamlining Redundant Layers to Compress Large Language ModelsCode1
Compacting, Picking and Growing for Unforgetting Continual LearningCode1
Composable Interventions for Language ModelsCode1
Contrastive Distillation on Intermediate Representations for Language Model CompressionCode1
Communication-Efficient Diffusion Strategy for Performance Improvement of Federated Learning with Non-IID DataCode1
Communication-Computation Trade-Off in Resource-Constrained Edge InferenceCode1
Communication-Efficient Federated Learning through Adaptive Weight Clustering and Server-Side DistillationCode1
Model LEGO: Creating Models Like Disassembling and Assembling Building BlocksCode1
Clustered Sampling: Low-Variance and Improved Representativity for Clients Selection in Federated LearningCode1
Show:102550
← PrevPage 3 of 55Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified