SOTAVerified

Network Pruning

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Papers

Showing 151200 of 534 papers

TitleStatusHype
Hyperflows: Pruning Reveals the Importance of Weights0
Can network pruning benefit deep learning under label noise?0
Energy Consumption of Neural Networks on NVIDIA Edge Boards: an Empirical Model0
C2S2: Cost-aware Channel Sparse Selection for Progressive Network Pruning0
Enabling Image Recognition on Constrained Devices Using Neural Network Pruning and a CycleGAN0
Brain-Inspired Efficient Pruning: Exploiting Criticality in Spiking Neural Networks0
A Novel Architecture Slimming Method for Network Pruning and Knowledge Distillation0
Hyperparameter Optimization with Neural Network Pruning0
LEAN: graph-based pruning for convolutional neural networks by extracting longest chains0
A "Network Pruning Network" Approach to Deep Model Compression0
Efficient Multi-Object Tracking on Edge Devices via Reconstruction-Based Channel Pruning0
Explicit Group Sparse Projection with Applications to Deep Learning and NMF0
Bi-LSTM based Multi-Agent DRL with Computation-aware Pruning for Agent Twins Migration in Vehicular Embodied AI Networks0
Efficient Ensembles of Graph Neural Networks0
Effective Subset Selection Through The Lens of Neural Network Pruning0
Graph Attention Network based Pruning for Reconstructing 3D Liver Vessel Morphology from Contrasted CT Images0
Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures0
An End-to-End Network Pruning Pipeline with Sparsity Enforcement0
GPU Acceleration of Sparse Neural Networks0
Dynamic parameter reallocation improves trainability of deep convolutional networks0
Accelerating Convolutional Neural Network Pruning via Spatial Aura Entropy0
GraFT: Gradual Fusion Transformer for Multimodal Re-Identification0
Hierarchical Action Classification with Network Pruning0
Dynamic ASR Pathways: An Adaptive Masking Approach Towards Efficient Pruning of A Multilingual ASR Model0
Bayesian Neural Networks at Scale: A Performance Analysis and Pruning Study0
Dynamic Probabilistic Pruning: A general framework for hardware-constrained pruning at different granularities0
Dynamical Isometry: The Missing Ingredient for Neural Network Pruning0
Basis Scaling and Double Pruning for Efficient Inference in Network-Based Transfer Learning0
A Multi-objective Complex Network Pruning Framework Based on Divide-and-conquer and Global Performance Impairment Ranking0
Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices0
EdgeBERT: Sentence-Level Energy Optimizations for Latency-Aware Multi-Task NLP Inference0
Diversity-boosted Generalization-Specialization Balancing for Zero-shot Learning0
CCSRP: Robust Pruning of Spiking Neural Networks through Cooperative Coevolution0
Does a sparse ReLU network training problem always admit an optimum?0
Blending Pruning Criteria for Convolutional Neural Networks0
Distortion-Aware Network Pruning and Feature Reuse for Real-time Video Segmentation0
An Efficient Row-Based Sparse Fine-Tuning0
GDP: Stabilized Neural Network Pruning via Gates with Differentiable Polarization0
Block Pruning for Enhanced Efficiency in Convolutional Neural Networks0
Efficient Stein Variational Inference for Reliable Distribution-lossless Network Pruning0
Analysing Neural Network Topologies: a Game Theoretic Approach0
Finding Stable Subnetworks at Initialization with Dataset Distillation0
Pruning-Aware Merging for Efficient Multitask Inference0
Discriminative Adversarial Unlearning0
AutoPruning for Deep Neural Network with Dynamic Channel Masking0
Enhanced Exploration in Neural Feature Selection for Deep Click-Through Rate Prediction Models via Ensemble of Gating Layers0
Enhancing Scalability in Recommender Systems through Lottery Ticket Hypothesis and Knowledge Distillation-based Neural Network Pruning0
Ensemble Mask Networks0
Finding Task-specific Subnetworks in Multi-task Spoken Language Understanding Model0
DIPNet: Efficiency Distillation and Iterative Pruning for Image Super-Resolution0
Show:102550
← PrevPage 4 of 11Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50-2.3 GFLOPsAccuracy78.79Unverified
2ResNet50-1.5 GFLOPsAccuracy78.07Unverified
3ResNet50 2.5 GFLOPSAccuracy78Unverified
4RegX-1.6GAccuracy77.97Unverified
5ResNet50 2.0 GFLOPSAccuracy77.7Unverified
6ResNet50-3G FLOPsAccuracy77.1Unverified
7ResNet50-2G FLOPsAccuracy76.4Unverified
8ResNet50-1G FLOPsAccuracy76.38Unverified
9TAS-pruned ResNet-50Accuracy76.2Unverified
10ResNet50Accuracy75.59Unverified
#ModelMetricClaimedVerifiedStatus
1FeatherTop-1 Accuracy76.93Unverified
2SpartanTop-1 Accuracy76.17Unverified
3ST-3Top-1 Accuracy76.03Unverified
4AC/DCTop-1 Accuracy75.64Unverified
5CSTop-1 Accuracy75.5Unverified
6ProbMaskTop-1 Accuracy74.68Unverified
7STRTop-1 Accuracy74.31Unverified
8DNWTop-1 Accuracy74Unverified
9GMPTop-1 Accuracy73.91Unverified
#ModelMetricClaimedVerifiedStatus
1+U-DML*Inference Time (ms)675.56Unverified
2DenseAccuracy79Unverified
3AC/DCAccuracy78.2Unverified
4Beta-RankAccuracy74.01Unverified
5TAS-pruned ResNet-110Accuracy73.16Unverified
#ModelMetricClaimedVerifiedStatus
1TAS-pruned ResNet-110Accuracy94.33Unverified
2ShuffleNet – QuantisedInference Time (ms)23.15Unverified
3AlexNet – QuantisedInference Time (ms)5.23Unverified
4MobileNet – QuantisedInference Time (ms)4.74Unverified
#ModelMetricClaimedVerifiedStatus
1FFN-ShapleyPrunedAvg #Steps12.05Unverified