SOTAVerified

Model Compression

Model Compression is an actively pursued area of research over the last few years with the goal of deploying state-of-the-art deep networks in low-power and resource limited devices without significant drop in accuracy. Parameter pruning, low-rank factorization and weight quantization are some of the proposed methods to compress the size of deep networks.

Source: KD-MRI: A knowledge distillation framework for image reconstruction and image restoration in MRI workflow

Papers

Showing 661670 of 1356 papers

TitleStatusHype
Interpreting Deep Classifier by Visual Distillation of Dark Knowledge0
Redundancy and Concept Analysis for Code-trained Language Models0
Intrinsically Sparse Long Short-Term Memory Networks0
Investigation of Practical Aspects of Single Channel Speech Separation for ASR0
Data-Model-Circuit Tri-Design for Ultra-Light Video Intelligence on Edge Devices0
Is Quantum Optimization Ready? An Effort Towards Neural Network Compression using Adiabatic Quantum Computing0
IteRABRe: Iterative Recovery-Aided Block Reduction0
Iterative Compression of End-to-End ASR Model using AutoML0
Decoupling Weight Regularization from Batch Size for Model Compression0
FinGPT-HPC: Efficient Pretraining and Finetuning Large Language Models for Financial Applications with High-Performance Computing0
Show:102550
← PrevPage 67 of 136Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MobileBERT + 2bit-1dim model compression using DKMAccuracy82.13Unverified
2MobileBERT + 1bit-1dim model compression using DKMAccuracy63.17Unverified