Statistical language models within the algebra of weighted rational languages
AbstractStatistical language models are an important tool in natural language processing. They represent prior knowledge about a certain language which is usually gained from a set of samples called a corpus. In this paper, we present a novel way of creating N-gram language models using weighted finite automata. The construction of these models is formalised within the algebra underlying weighted finite automata and expressed in terms of weighted rational languages and transductions. Besides the algebra we make use of five special constant weighted transductions which rely only on the alphabet and the model parameter N. In addition, we discuss efficient implementations of these transductions in terms of virtual constructions.
Download data is not yet available.
How to Cite
Hanneforth, T., & Würzner, K.-M. (2009). Statistical language models within the algebra of weighted rational languages. Acta Cybernetica, 19(2), 313-356. Retrieved from https://cyber.bibl.u-szeged.hu/index.php/actcybern/article/view/3771