SOTAVerified

Hierarchical Quantized Autoencoders

2020-02-19NeurIPS 2020Code Available1· sign in to hype

Will Williams, Sam Ringer, Tom Ash, John Hughes, David MacLeod, Jamie Dougherty

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Despite progress in training neural networks for lossy image compression, current approaches fail to maintain both perceptual quality and abstract features at very low bitrates. Encouraged by recent success in learning discrete representations with Vector Quantized Variational Autoencoders (VQ-VAEs), we motivate the use of a hierarchy of VQ-VAEs to attain high factors of compression. We show that the combination of stochastic quantization and hierarchical latent structure aids likelihood-based image compression. This leads us to introduce a novel objective for training hierarchical VQ-VAEs. Our resulting scheme produces a Markovian series of latent variables that reconstruct images of high-perceptual quality which retain semantically meaningful features. We provide qualitative and quantitative evaluations on the CelebA and MNIST datasets.

Tasks

Reproductions