SOTAVerified

Neural Text Normalization with Subword Units

2019-06-01NAACL 2019Unverified0· sign in to hype

Courtney Mansfield, Ming Sun, Yuzong Liu, G, Ankur he, Bj{\"o}rn Hoffmeister

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Text normalization (TN) is an important step in conversational systems. It converts written text to its spoken form to facilitate speech recognition, natural language understanding and text-to-speech synthesis. Finite state transducers (FSTs) are commonly used to build grammars that handle text normalization. However, translating linguistic knowledge into grammars requires extensive effort. In this paper, we frame TN as a machine translation task and tackle it with sequence-to-sequence (seq2seq) models. Previous research focuses on normalizing a word (or phrase) with the help of limited word-level context, while our approach directly normalizes full sentences. We find subword models with additional linguistic features yield the best performance (with a word error rate of 0.17\%).

Tasks

Reproductions