Javier Gónzalez-DomínguezDavid EustisIgnacio López MorenoAndrew SeniorFrançoise BeaufaysPedro J. Moreno
Automatic speech recognition (ASR) systems are used daily by millions of people worldwide to dictate messages, control devices, initiate searches or to facilitate data input in small devices. The user experience in these scenarios depends on the quality of the speech transcriptions and on the responsiveness of the system. For multilingual users, a further obstacle to natural interaction is the monolingual character of many ASR systems, in which users are constrained to a single preset language. In this work, we present an end-to-end multi-language ASR architecture, developed and deployed at Google, that allows users to select arbitrary combinations of spoken languages. We leverage recent advances in language identification and a novel method of real-time language selection to achieve similar recognition accuracy and nearly-identical latency characteristics as a monolingual system.
Hiroshi SekiTakaaki HoriShinji WatanabeJonathan Le RouxJohn R. Hershey
Li SongBeibei OuyangFuchuan TongDexin LiaoLin LiQingyang Hong
Austin WatersNeeraj GaurParisa HaghaniPedro J. MorenoZhongdi Qu
Shubham ToshniwalTara N. SainathRon J. WeissBo LiPedro J. MorenoEugene WeinsteinKanishka Rao
Hirofumi InagumaKevin DuhTatsuya KawaharaShinji Watanabe