SOTAVerified

Network Pruning

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Papers

Showing 451500 of 534 papers

TitleStatusHype
Waste not, Want not: All-Alive Pruning for Extremely Sparse Networks0
Weight-dependent Gates for Network Pruning0
Understanding the Effects of Data Parallelism and Sparsity on Neural Network Training0
Connection Sensitivity Matters for Training-free DARTS: From Architecture-Level Scoring to Operation-Level Sensitivity Analysis0
Differentiable Channel Sparsity Search via Weight Sharing within Filters0
Differentiable Network Pruning for Microcontrollers0
Differential Privacy Meets Neural Network Pruning0
Selective Brain Damage: Measuring the Disparate Impact of Model Pruning0
DIPNet: Efficiency Distillation and Iterative Pruning for Image Super-Resolution0
Data Augmentations in Deep Weight Spaces0
Discriminative Adversarial Unlearning0
Pruning-Aware Merging for Efficient Multitask Inference0
Self-Adaptive Network Pruning0
Analysing Neural Network Topologies: a Game Theoretic Approach0
Distortion-Aware Network Pruning and Feature Reuse for Real-time Video Segmentation0
Does a sparse ReLU network training problem always admit an optimum?0
SequentialAttention++ for Block Sparsification: Differentiable Pruning Meets Combinatorial Optimization0
DARB: A Density-Aware Regular-Block Pruning for Deep Neural Networks0
A Main/Subsidiary Network Framework for Simplifying Binary Neural Network0
Dynamical Isometry: The Missing Ingredient for Neural Network Pruning0
Dynamic ASR Pathways: An Adaptive Masking Approach Towards Efficient Pruning of A Multilingual ASR Model0
dagger: A Python Framework for Reproducible Machine Learning Experiment Orchestration0
Dynamic parameter reallocation improves trainability of deep convolutional networks0
Dynamic Probabilistic Pruning: A general framework for hardware-constrained pruning at different granularities0
Cross-Channel Intragroup Sparsity Neural Network0
Crossbar-aware neural network pruning0
Convolutional Neural Network Pruning with Structural Redundancy Reduction0
Signal Collapse in One-Shot Pruning: When Sparse Models Fail to Distinguish Neural Representations0
EdgeBERT: Sentence-Level Energy Optimizations for Latency-Aware Multi-Task NLP Inference0
Convolutional Neural Network Pruning Using Filter Attenuation0
Effective Subset Selection Through The Lens of Neural Network Pruning0
Efficient Ensembles of Graph Neural Networks0
A Lottery Ticket Hypothesis Framework for Low-Complexity Device-Robust Neural Acoustic Scene Classification0
Convolutional Neural Network Pruning to Accelerate Membrane Segmentation in Electron Microscopy0
ConfusionFlow: A model-agnostic visualization for temporal analysis of classifier confusion0
Efficient Multi-Object Tracking on Edge Devices via Reconstruction-Based Channel Pruning0
Single-shot Channel Pruning Based on Alternating Direction Method of Multipliers0
Efficient Stein Variational Inference for Reliable Distribution-lossless Network Pruning0
Confident magnitude-based neural network pruning0
Small Contributions, Small Networks: Efficient Neural Network Pruning Based on Relative Importance0
Concept-Monitor: Understanding DNN training through individual neurons0
Enabling Image Recognition on Constrained Devices Using Neural Network Pruning and a CycleGAN0
Energy Consumption of Neural Networks on NVIDIA Edge Boards: an Empirical Model0
Enhanced Exploration in Neural Feature Selection for Deep Click-Through Rate Prediction Models via Ensemble of Gating Layers0
Enhancing Scalability in Recommender Systems through Lottery Ticket Hypothesis and Knowledge Distillation-based Neural Network Pruning0
Ensemble Mask Networks0
EPSD: Early Pruning with Self-Distillation for Efficient Model Compression0
SMOF: Squeezing More Out of Filters Yields Hardware-Friendly CNN Pruning0
ExPAN(N)D: Exploring Posits for Efficient Artificial Neural Network Design in FPGA-based Systems0
Improving Feature Attribution through Input-specific Network Pruning0
Show:102550
← PrevPage 10 of 11Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50-2.3 GFLOPsAccuracy78.79Unverified
2ResNet50-1.5 GFLOPsAccuracy78.07Unverified
3ResNet50 2.5 GFLOPSAccuracy78Unverified
4RegX-1.6GAccuracy77.97Unverified
5ResNet50 2.0 GFLOPSAccuracy77.7Unverified
6ResNet50-3G FLOPsAccuracy77.1Unverified
7ResNet50-2G FLOPsAccuracy76.4Unverified
8ResNet50-1G FLOPsAccuracy76.38Unverified
9TAS-pruned ResNet-50Accuracy76.2Unverified
10ResNet50Accuracy75.59Unverified
#ModelMetricClaimedVerifiedStatus
1FeatherTop-1 Accuracy76.93Unverified
2SpartanTop-1 Accuracy76.17Unverified
3ST-3Top-1 Accuracy76.03Unverified
4AC/DCTop-1 Accuracy75.64Unverified
5CSTop-1 Accuracy75.5Unverified
6ProbMaskTop-1 Accuracy74.68Unverified
7STRTop-1 Accuracy74.31Unverified
8DNWTop-1 Accuracy74Unverified
9GMPTop-1 Accuracy73.91Unverified
#ModelMetricClaimedVerifiedStatus
1+U-DML*Inference Time (ms)675.56Unverified
2DenseAccuracy79Unverified
3AC/DCAccuracy78.2Unverified
4Beta-RankAccuracy74.01Unverified
5TAS-pruned ResNet-110Accuracy73.16Unverified
#ModelMetricClaimedVerifiedStatus
1TAS-pruned ResNet-110Accuracy94.33Unverified
2ShuffleNet – QuantisedInference Time (ms)23.15Unverified
3AlexNet – QuantisedInference Time (ms)5.23Unverified
4MobileNet – QuantisedInference Time (ms)4.74Unverified
#ModelMetricClaimedVerifiedStatus
1FFN-ShapleyPrunedAvg #Steps12.05Unverified