SOTAVerified

AntMan: Sparse Low-Rank Compression to Accelerate RNN inference

2019-10-02ICLR 2019Unverified0· sign in to hype

Samyam Rajbhandari, Harsh Shrivastava, Yuxiong He

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Wide adoption of complex RNN based models is hindered by their inference performance, cost and memory requirements. To address this issue, we develop AntMan, combining structured sparsity with low-rank decomposition synergistically, to reduce model computation, size and execution time of RNNs while attaining desired accuracy. AntMan extends knowledge distillation based training to learn the compressed models efficiently. Our evaluation shows that AntMan offers up to 100x computation reduction with less than 1pt accuracy drop for language and machine reading comprehension models. Our evaluation also shows that for a given accuracy target, AntMan produces 5x smaller models than the state-of-art. Lastly, we show that AntMan offers super-linear speed gains compared to theoretical speedup, demonstrating its practical value on commodity hardware.

Tasks

Reproductions