SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 321330 of 1356 papers

TitleStatusHype
Bayesian Federated Model Compression for Communication and Computation Efficiency0
Bayesian Deep Learning Via Expectation Maximization and Turbo Deep Approximate Message Passing0
A Model Compression Method with Matrix Product Operators for Speech Enhancement0
Design and Prototyping Distributed CNN Inference Acceleration in Edge Computing0
A Mixed Integer Programming Approach for Verifying Properties of Binarized Neural Networks0
Balancing Specialization, Generalization, and Compression for Detection and Tracking0
Balancing Cost and Benefit with Tied-Multi Transformers0
Activation Map Adaptation for Effective Knowledge Distillation0
Single-path Bit Sharing for Automatic Loss-aware Model Compression0
Design Automation for Fast, Lightweight, and Effective Deep Learning Models: A Survey0
Show:102550
← PrevPage 33 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified