SOTAVerified

Network Pruning

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Papers

Showing 301325 of 534 papers

TitleStatusHype
Bi-LSTM based Multi-Agent DRL with Computation-aware Pruning for Agent Twins Migration in Vehicular Embodied AI Networks0
Surrogate Lagrangian Relaxation: A Path To Retrain-free Deep Neural Network Pruning0
Maximum Redundancy Pruning: A Principle-Driven Layerwise Sparsity Allocation for LLMs0
Mechanistically analyzing the effects of fine-tuning on procedurally defined tasks0
Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures0
Mini-batch Coresets for Memory-efficient Training of Large Language Models0
Meta-Learning with Network Pruning0
Meta-Learning with Network Pruning for Overfitting Reduction0
Bayesian Neural Networks at Scale: A Performance Analysis and Pruning Study0
The Generalization-Stability Tradeoff In Neural Network Pruning0
Model compression as constrained optimization, with application to neural nets. Part V: combining compressions0
Model Compression Methods for YOLOv5: A Review0
Model Pruning Enables Localized and Efficient Federated Learning for Yield Forecasting and Data Sharing0
Basis Scaling and Double Pruning for Efficient Inference in Network-Based Transfer Learning0
Multi-Agent Actor-Critic with Harmonic Annealing Pruning for Dynamic Spectrum Access Systems0
Multistage Pruning of CNN Based ECG Classifiers for Edge Devices0
Multi-Task Network Pruning and Embedded Optimization for Real-time Deployment in ADAS0
Mutual Information Preserving Neural Network Pruning0
N2NSkip: Learning Highly Sparse Networks using Neuron-to-Neuron Skip Connections0
NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm0
Diversity-boosted Generalization-Specialization Balancing for Zero-shot Learning0
Network Automatic Pruning: Start NAP and Take a Nap0
The Incredible Shrinking Neural Network: New Perspectives on Learning Representations Through The Lens of Pruning0
Network Pruning by Greedy Subnetwork Selection0
Network Pruning for Low-Rank Binary Index0
Show:102550
← PrevPage 13 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50-2.3 GFLOPsAccuracy78.79Unverified
2ResNet50-1.5 GFLOPsAccuracy78.07Unverified
3ResNet50 2.5 GFLOPSAccuracy78Unverified
4RegX-1.6GAccuracy77.97Unverified
5ResNet50 2.0 GFLOPSAccuracy77.7Unverified
6ResNet50-3G FLOPsAccuracy77.1Unverified
7ResNet50-2G FLOPsAccuracy76.4Unverified
8ResNet50-1G FLOPsAccuracy76.38Unverified
9TAS-pruned ResNet-50Accuracy76.2Unverified
10ResNet50Accuracy75.59Unverified
#ModelMetricClaimedVerifiedStatus
1FeatherTop-1 Accuracy76.93Unverified
2SpartanTop-1 Accuracy76.17Unverified
3ST-3Top-1 Accuracy76.03Unverified
4AC/DCTop-1 Accuracy75.64Unverified
5CSTop-1 Accuracy75.5Unverified
6ProbMaskTop-1 Accuracy74.68Unverified
7STRTop-1 Accuracy74.31Unverified
8DNWTop-1 Accuracy74Unverified
9GMPTop-1 Accuracy73.91Unverified
#ModelMetricClaimedVerifiedStatus
1+U-DML*Inference Time (ms)675.56Unverified
2DenseAccuracy79Unverified
3AC/DCAccuracy78.2Unverified
4Beta-RankAccuracy74.01Unverified
5TAS-pruned ResNet-110Accuracy73.16Unverified
#ModelMetricClaimedVerifiedStatus
1TAS-pruned ResNet-110Accuracy94.33Unverified
2ShuffleNet – QuantisedInference Time (ms)23.15Unverified
3AlexNet – QuantisedInference Time (ms)5.23Unverified
4MobileNet – QuantisedInference Time (ms)4.74Unverified
#ModelMetricClaimedVerifiedStatus
1FFN-ShapleyPrunedAvg #Steps12.05Unverified