SOTAVerified

Speaker-Independent Acoustic-to-Articulatory Inversion through Multi-Channel Attention Discriminator

2024-06-25Code Available0· sign in to hype

Woo-Jin Chung, Hong-Goo Kang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present a novel speaker-independent acoustic-to-articulatory inversion (AAI) model, overcoming the limitations observed in conventional AAI models that rely on acoustic features derived from restricted datasets. To address these challenges, we leverage representations from a pre-trained self-supervised learning (SSL) model to more effectively estimate the global, local, and kinematic pattern information in Electromagnetic Articulography (EMA) signals during the AAI process. We train our model using an adversarial approach and introduce an attention-based Multi-duration phoneme discriminator (MDPD) designed to fully capture the intricate relationship among multi-channel articulatory signals. Our method achieves a Pearson correlation coefficient of 0.847, marking state-of-the-art performance in speaker-independent AAI models. The implementation details and code can be found online.

Tasks

Reproductions