SOTAVerified

Network Pruning

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Papers

Showing 451475 of 534 papers

TitleStatusHype
Cross-Channel Intragroup Sparsity Neural Network0
Self-Adaptive Network Pruning0
Privacy Preserving Stochastic Channel-Based Federated Learning with Neural Network Pruning0
ConfusionFlow: A model-agnostic visualization for temporal analysis of classifier confusion0
Pruning from ScratchCode0
Network Pruning for Low-Rank Binary Index0
Selective Brain Damage: Measuring the Disparate Impact of Model Pruning0
Finding Deep Local Optima Using Network Pruning0
Meta-Learning with Network Pruning for Overfitting Reduction0
On the Decision Boundaries of Deep Neural Networks: A Tropical Geometry Perspective0
Class-dependent Compression of Deep Neural NetworksCode0
Defeating Misclassification Attacks Against Transfer Learning0
Architecture-aware Network Pruning for Vision Quality Applications0
A Targeted Acceleration and Compression Framework for Low bit Neural Networks0
Data-Independent Neural Pruning via Coresets0
Importance Estimation for Neural Network PruningCode0
Joint Regularization on Activations and Weights for Efficient Neural Network Pruning0
Structured Pruning of Recurrent Neural Networks through Neuron Selection0
Towards Compact and Robust Deep Neural Networks0
A Signal Propagation Perspective for Pruning Neural Networks at InitializationCode0
The Generalization-Stability Tradeoff In Neural Network Pruning0
Discovering Neural WiringsCode1
Variational Convolutional Neural Network Pruning0
Learning Sparse Networks Using Targeted DropoutCode0
Pruning-Aware Merging for Efficient Multitask Inference0
Show:102550
← PrevPage 19 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50-2.3 GFLOPsAccuracy78.79Unverified
2ResNet50-1.5 GFLOPsAccuracy78.07Unverified
3ResNet50 2.5 GFLOPSAccuracy78Unverified
4RegX-1.6GAccuracy77.97Unverified
5ResNet50 2.0 GFLOPSAccuracy77.7Unverified
6ResNet50-3G FLOPsAccuracy77.1Unverified
7ResNet50-2G FLOPsAccuracy76.4Unverified
8ResNet50-1G FLOPsAccuracy76.38Unverified
9TAS-pruned ResNet-50Accuracy76.2Unverified
10ResNet50Accuracy75.59Unverified
#ModelMetricClaimedVerifiedStatus
1FeatherTop-1 Accuracy76.93Unverified
2SpartanTop-1 Accuracy76.17Unverified
3ST-3Top-1 Accuracy76.03Unverified
4AC/DCTop-1 Accuracy75.64Unverified
5CSTop-1 Accuracy75.5Unverified
6ProbMaskTop-1 Accuracy74.68Unverified
7STRTop-1 Accuracy74.31Unverified
8DNWTop-1 Accuracy74Unverified
9GMPTop-1 Accuracy73.91Unverified
#ModelMetricClaimedVerifiedStatus
1+U-DML*Inference Time (ms)675.56Unverified
2DenseAccuracy79Unverified
3AC/DCAccuracy78.2Unverified
4Beta-RankAccuracy74.01Unverified
5TAS-pruned ResNet-110Accuracy73.16Unverified
#ModelMetricClaimedVerifiedStatus
1TAS-pruned ResNet-110Accuracy94.33Unverified
2ShuffleNet – QuantisedInference Time (ms)23.15Unverified
3AlexNet – QuantisedInference Time (ms)5.23Unverified
4MobileNet – QuantisedInference Time (ms)4.74Unverified
#ModelMetricClaimedVerifiedStatus
1FFN-ShapleyPrunedAvg #Steps12.05Unverified