SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 911920 of 1356 papers

TitleStatusHype
Pruning Ternary Quantization0
Accelerating deep neural networks for efficient scene understanding in automotive cyber-physical systems0
A New Clustering-Based Technique for the Acceleration of Deep Convolutional Networks0
Federated Action Recognition on Heterogeneous Embedded Devices0
Efficient automated U-Net based tree crown delineation using UAV multi-spectral imagery on embedded devices0
Compact and Optimal Deep Learning with Recurrent Parameter GeneratorsCode0
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions0
WeClick: Weakly-Supervised Video Semantic Segmentation with Click Annotations0
Universal approximation and model compression for radial neural networksCode0
A Light-weight Deep Human Activity Recognition Algorithm Using Multi-knowledge Distillation0
Show:102550
← PrevPage 92 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified