Scalable Multi Corpora Neural Language Models for ASR

07/02/2019
by   Anirudh Raju, et al.
0

Neural language models (NLM) have been shown to outperform conventional n-gram language models by a substantial margin in Automatic Speech Recognition (ASR) and other tasks. There are, however, a number of challenges that need to be addressed for an NLM to be used in a practical large-scale ASR system. In this paper, we present solutions to some of the challenges, including training NLM from heterogenous corpora, limiting latency impact and handling personalized bias in the second-pass rescorer. Overall, we show that we can achieve a 6.2 rescoring framework with a minimal increase in latency.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset