Factorization tricks for LSTM networks
2017-03-31Code Available1· sign in to hype
Oleksii Kuchaiev, Boris Ginsburg
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/okuchaiev/f-lmOfficialIn papertf★ 0
- github.com/rdspring1/PyTorch_GBW_LMpytorch★ 123
Abstract
We present two simple ways of reducing the number of parameters and accelerating the training of large Long Short-Term Memory (LSTM) networks: the first one is "matrix factorization by design" of LSTM matrix into the product of two smaller matrices, and the second one is partitioning of LSTM matrix, its inputs and states into the independent groups. Both approaches allow us to train large LSTM networks significantly faster to the near state-of the art perplexity while using significantly less RNN parameters.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| One Billion Word | BIG G-LSTM-2 | PPL | 36 | — | Unverified |