SOTAVerified

Network Pruning

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Papers

Showing 451500 of 534 papers

TitleStatusHype
Quantum-Inspired Hamiltonian Monte Carlo for Bayesian SamplingCode0
AutoPrune: Automatic Network Pruning by Regularizing Auxiliary ParametersCode0
One-Shot Pruning of Recurrent Neural Networks by Jacobian Spectrum Evaluation0
Improving Feature Attribution through Input-specific Network Pruning0
Neural Network Pruning with Residual-Connections and Limited-DataCode0
DARB: A Density-Aware Regular-Block Pruning for Deep Neural Networks0
What Do Compressed Deep Neural Networks Forget?Code0
Iteratively Training Look-Up Tables for Network Quantization0
Cross-Channel Intragroup Sparsity Neural Network0
Self-Adaptive Network Pruning0
Privacy Preserving Stochastic Channel-Based Federated Learning with Neural Network Pruning0
ConfusionFlow: A model-agnostic visualization for temporal analysis of classifier confusion0
Pruning from ScratchCode0
Network Pruning for Low-Rank Binary Index0
On the Decision Boundaries of Deep Neural Networks: A Tropical Geometry Perspective0
Finding Deep Local Optima Using Network Pruning0
Selective Brain Damage: Measuring the Disparate Impact of Model Pruning0
Meta-Learning with Network Pruning for Overfitting Reduction0
Class-dependent Compression of Deep Neural NetworksCode0
Defeating Misclassification Attacks Against Transfer Learning0
Architecture-aware Network Pruning for Vision Quality Applications0
Data-Independent Neural Pruning via Coresets0
A Targeted Acceleration and Compression Framework for Low bit Neural Networks0
Importance Estimation for Neural Network PruningCode0
Joint Regularization on Activations and Weights for Efficient Neural Network Pruning0
Structured Pruning of Recurrent Neural Networks through Neuron Selection0
A Signal Propagation Perspective for Pruning Neural Networks at InitializationCode0
Towards Compact and Robust Deep Neural Networks0
The Generalization-Stability Tradeoff In Neural Network Pruning0
Variational Convolutional Neural Network Pruning0
Learning Sparse Networks Using Targeted DropoutCode0
Network Pruning via Transformable Architecture SearchCode0
Pruning-Aware Merging for Efficient Multitask Inference0
EigenDamage: Structured Pruning in the Kronecker-Factored EigenbasisCode0
Network Pruning for Low-Rank Binary Indexing0
Towards Learning of Filter-Level Heterogeneous Compression of Convolutional Neural NetworksCode0
Analysing Neural Network Topologies: a Game Theoretic Approach0
C2S2: Cost-aware Channel Sparse Selection for Progressive Network Pruning0
Progressive Stochastic Binarization of Deep NetworksCode0
Adversarial Robustness vs Model Compression, or Both?Code0
FastDepth: Fast Monocular Depth Estimation on Embedded SystemsCode0
Single-shot Channel Pruning Based on Alternating Direction Method of Multipliers0
Same, Same But Different - Recovering Neural Network Quantization Error Through Weight FactorizationCode0
DeepSZ: A Novel Framework to Compress Deep Neural Networks by Using Error-Bounded Lossy CompressionCode0
GASL: Guided Attention for Sparsity Learning in Deep Neural NetworksCode0
A Main/Subsidiary Network Framework for Simplifying Binary Neural Network0
Few Sample Knowledge Distillation for Efficient Network CompressionCode0
Channel-wise pruning of neural networks with tapering resource constraint0
Neural Rejuvenation: Improving Deep Network Training by Enhancing Computational Resource UtilizationCode0
Efficient Structured Pruning and Architecture Searching for Group ConvolutionCode0
Show:102550
← PrevPage 10 of 11Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50-2.3 GFLOPsAccuracy78.79Unverified
2ResNet50-1.5 GFLOPsAccuracy78.07Unverified
3ResNet50 2.5 GFLOPSAccuracy78Unverified
4RegX-1.6GAccuracy77.97Unverified
5ResNet50 2.0 GFLOPSAccuracy77.7Unverified
6ResNet50-3G FLOPsAccuracy77.1Unverified
7ResNet50-2G FLOPsAccuracy76.4Unverified
8ResNet50-1G FLOPsAccuracy76.38Unverified
9TAS-pruned ResNet-50Accuracy76.2Unverified
10ResNet50Accuracy75.59Unverified
#ModelMetricClaimedVerifiedStatus
1FeatherTop-1 Accuracy76.93Unverified
2SpartanTop-1 Accuracy76.17Unverified
3ST-3Top-1 Accuracy76.03Unverified
4AC/DCTop-1 Accuracy75.64Unverified
5CSTop-1 Accuracy75.5Unverified
6ProbMaskTop-1 Accuracy74.68Unverified
7STRTop-1 Accuracy74.31Unverified
8DNWTop-1 Accuracy74Unverified
9GMPTop-1 Accuracy73.91Unverified
#ModelMetricClaimedVerifiedStatus
1+U-DML*Inference Time (ms)675.56Unverified
2DenseAccuracy79Unverified
3AC/DCAccuracy78.2Unverified
4Beta-RankAccuracy74.01Unverified
5TAS-pruned ResNet-110Accuracy73.16Unverified
#ModelMetricClaimedVerifiedStatus
1TAS-pruned ResNet-110Accuracy94.33Unverified
2ShuffleNet – QuantisedInference Time (ms)23.15Unverified
3AlexNet – QuantisedInference Time (ms)5.23Unverified
4MobileNet – QuantisedInference Time (ms)4.74Unverified
#ModelMetricClaimedVerifiedStatus
1FFN-ShapleyPrunedAvg #Steps12.05Unverified