Parallel WaveGAN: A fast waveform generation model based on generative adversarial networks with multi-resolution spectrogram
Ryuichi Yamamoto, Eunwoo Song, Jae-Min Kim
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/coqui-ai/TTSpytorch★ 44,894
- github.com/PaddlePaddle/PaddleSpeechpaddle★ 12,564
- github.com/TensorSpeech/TensorflowTTStf★ 3,995
- github.com/facebookresearch/denoiserpytorch★ 1,882
- github.com/bigpon/vcc20_baseline_cyclevaepytorch★ 131
- github.com/bigpon/QPPWGpytorch★ 46
- github.com/yanggeng1995/GAN-TTSpytorch★ 0
- github.com/Moon-sung-woo/ParallelWaveGan_koreanpytorch★ 0
- github.com/yanggeng1995/FB-MelGANpytorch★ 0
- github.com/mukeshv0/ParallelWaveGANpytorch★ 0
Abstract
We propose Parallel WaveGAN, a distillation-free, fast, and small-footprint waveform generation method using a generative adversarial network. In the proposed method, a non-autoregressive WaveNet is trained by jointly optimizing multi-resolution spectrogram and adversarial loss functions, which can effectively capture the time-frequency distribution of the realistic speech waveform. As our method does not require density distillation used in the conventional teacher-student framework, the entire model can be easily trained. Furthermore, our model is able to generate high-fidelity speech even with its compact architecture. In particular, the proposed Parallel WaveGAN has only 1.44 M parameters and can generate 24 kHz speech waveform 28.68 times faster than real-time on a single GPU environment. Perceptual listening test results verify that our proposed method achieves 4.16 mean opinion score within a Transformer-based text-to-speech framework, which is comparative to the best distillation-based Parallel WaveNet system.