SOTAVerified

Training and Generating Neural Networks in Compressed Weight Space

2021-12-31ICLR Workshop Neural_Compression 2021Code Available0· sign in to hype

Kazuki Irie, Jürgen Schmidhuber

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The inputs and/or outputs of some neural nets are weight matrices of other neural nets. Indirect encodings or end-to-end compression of weight matrices could help to scale such approaches. Our goal is to open a discussion on this topic, starting with recurrent neural networks for character-level language modelling whose weight matrices are encoded by the discrete cosine transform. Our fast weight version thereof uses a recurrent neural network to parameterise the compressed weights. We present experimental results on the enwik8 dataset.

Tasks

Reproductions