Result Details
Approximate inference: A sampling based modeling technique to capture complex dependencies in a language model
Mikolov Tomáš, Ing., Ph.D., FIT (FIT), DCGM (FIT)
Kombrink Stefan, Dipl.-Linguist., FIT (FIT)
Church Kenneth
This paper deals with approximate inference: a sampling based modeling technique to capture complex dependencies in a language model
Long-span language models; Recurrent neural networks; Speech recognition; Decoding
In this paper, we present strategies to incorporate long context information directly during the first pass decoding and also for the second pass lattice re-scoring in speech recognition systems. Long-span language models that capture complex syntactic and/or semantic information are seldom used in the first pass of large vocabulary continuous speech recognition systems due to the prohibitive increase in the size of the sentence-hypotheses search space. Typically, n-gram language models are used in the first pass to produce N-best lists, which are then re-scored using long-span models. Such a pipeline produces biased first pass output, resulting in sub-optimal performance during re-scoring. In this paper we show that computationally tractable variational approximations of the long-span and complex language models are a better choice than the standard n-gram model for the first pass decoding and also for lattice re-scoring.
@article{BUT97047,
author="Anoop {Deoras} and Tomáš {Mikolov} and Stefan {Kombrink} and Kenneth {Church}",
title="Approximate inference: A sampling based modeling technique to capture complex dependencies in a language model",
journal="SPEECH COMMUNICATION",
year="2012",
volume="2012",
number="8",
pages="1--16",
doi="10.1016/j.specom.2012.08.004",
issn="0167-6393",
url="http://www.sciencedirect.com/science/article/pii/S0167639312000969#"
}
Technologies of speech processing for efficient human-machine communication, TAČR, Program aplikovaného výzkumu a experimentálního vývoje ALFA, TA01011328, start: 2011-01-01, end: 2014-12-31, completed