SOTAVerified

Storage Efficient and Dynamic Flexible Runtime Channel Pruning via Deep Reinforcement Learning

2020-12-01NeurIPS 2020Unverified0· sign in to hype

Jianda Chen, Shangyu Chen, Sinno Jialin Pan

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this paper, we propose a deep reinforcement learning (DRL) based framework to efficiently perform runtime channel pruning on convolutional neural networks (CNNs). Our DRL-based framework aims to learn a pruning strategy to determine how many and which channels to be pruned in each convolutional layer, depending on each individual input instance at runtime. Unlike existing runtime pruning methods which require to store all channels parameters for inference, our framework can reduce parameters storage consumption by introducing a static pruning component. Comparison experimental results with existing runtime and static pruning methods on state-of-the-art CNNs demonstrate that our proposed framework is able to provide a tradeoff between dynamic flexibility and storage efficiency in runtime channel pruning.

Tasks

Reproductions