Publication Details
Approximate inference: A sampling based modeling technique to capture complex dependencies in a language model
Mikolov Tomáš, Ing. (DCGM FIT BUT)
Kombrink Stefan, Dipl.-Inf -Ling (DCGM FIT BUT)
Church Kenneth (JHU)
Long-span language models; Recurrent neural networks; Speech recognition; Decoding
This paper deals with approximate inference: a sampling based modeling technique to capture complex dependencies in a language model
In this paper, we present strategies to incorporate long context information directly during the first pass decoding and also for the second pass lattice re-scoring in speech recognition systems. Long-span language models that capture complex syntactic and/or semantic information are seldom used in the first pass of large vocabulary continuous speech recognition systems due to the prohibitive increase in the size of the sentence-hypotheses search space. Typically, n-gram language models are used in the first pass to produce N-best lists, which are then re-scored using long-span models. Such a pipeline produces biased first pass output, resulting in sub-optimal performance during re-scoring. In this paper we show that computationally tractable variational approximations of the long-span and complex language models are a better choice than the standard n-gram model for the first pass decoding and also for lattice re-scoring.
@ARTICLE{FITPUB10160, author = "Anoop Deoras and Tom\'{a}\v{s} Mikolov and Stefan Kombrink and Kenneth Church", title = "Approximate inference: A sampling based modeling technique to capture complex dependencies in a language model", pages = "1--16", booktitle = "Speech Communication", journal = "Speech Communication", volume = 2012, number = 8, year = 2012, publisher = "Elsevier Science", ISSN = "0167-6393", doi = "10.1016/j.specom.2012.08.004", language = "english", url = "https://www.fit.vut.cz/research/publication/10160" }