[an error occurred while processing this directive]

Minimally Supervised Text Normalization: Algorithms and Implementation

Text normalization refers to methods used to convert text input from written (e.g., "$2.50", "5:15pm") to spoken form ("two dollars and fifty cents", "five fifteen PM"), primarily, though not exclusively, as part of a text-to-speech synthesis engine. Unlike much of modern natural language processing, much of this process is currently accomplished using hand-written language-specific grammars rather than general-purpose machine learning engines, hindering our ability to scale. In the first half of the talk, I describe experiments on learning text normalization for cardinal and ordinal numbers (Gorman & Sproat 2016), which accounts for much of the complexity of these language-specific grammars. The first model I describe uses an end-to-end recurrent neural network. The second model, drawing inspiration from the linguistics literature, uses finite-state transducers constructed with a minimal amount of training data. While both models achieve near-perfect performance, the latter model can be trained using several orders of magnitude less data than the former, making it particularly useful for low-resource languages. I also describe some extensions of this work to other classes such as currency and time expressions. The second half of the talk will discuss Pynini (Gorman 2016), an open-source Python library for grammar compilation, which is used to implement minimally supervised text normalization grammar induction. I will review the design of this library and provide several tutorial examples drawn from morphophonology. 

For more information about Kyle Gorman, click https://research.google.com/pubs/KyleGorman.html

Bio

Kyle Gorman is a computational linguist working on speech and language processing at Google. Before joining Google, he was a postdoctoral researcher, and an assistant professor, at the Center for Spoken Language Understanding at the Oregon Health & Science University. I received a Ph.D. in linguistics from the University of Pennsylvania in 2013. At Google, he contributes to the OpenFst and OpenGrm libraries, and is the principal author of Pynini, a powerful weighted-finite state grammar extension for Python. In his copious free time, Kyle also participates in ongoing collaborations in linguistics, language acquisition, and language disorders.

Speaker

Kyle Gorman, Google Research, NYC

Date

Wednesday, November 29, 2017

Time

4 - 6pm

Location

IACS Seminar Room