Generating Sequences With Recurrent Neural Networks
2013-08-04Code Available1· sign in to hype
Alex Graves
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/altsoph/paranoid_transformerpytorch★ 19
- github.com/robertknight/textgenpytorch★ 0
- github.com/CambridgeIIS/Gesture-Keyboard-Traj-Gentf★ 0
- github.com/hardmaru/sketch-rnn-datasetstf★ 0
- github.com/AnesBenmerzoug/Handwriting-Modelpytorch★ 0
- github.com/mohitRohatgi/handwritingGeneratornone★ 0
- github.com/YukiDaSlayer316/R-Tutorialnone★ 0
- github.com/sotelo/scribenone★ 0
- github.com/feiwu77777/Handwriting_generationtf★ 0
- github.com/raymondhs/char-rnn-truecasetorch★ 0
Abstract
This paper shows how Long Short-term Memory recurrent neural networks can be used to generate complex sequences with long-range structure, simply by predicting one data point at a time. The approach is demonstrated for text (where the data are discrete) and online handwriting (where the data are real-valued). It is then extended to handwriting synthesis by allowing the network to condition its predictions on a text sequence. The resulting system is able to generate highly realistic cursive handwriting in a wide variety of styles.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| enwik8 | LSTM (7 layers) | Bit per Character (BPC) | 1.67 | — | Unverified |