SuperLectures.com

VARIATIONAL APPROXIMATION OF LONG-SPAN LANGUAGE MODELS FOR LVCSR

Full Paper at IEEE Xplore

Language Modeling

Přednášející: Sanjeev Khudanpur, Autoři: Anoop Deoras, Center for Language and Speech Processing, United States; Tomáš Mikolov, Stefan Kombrink, Martin Karafiát, Brno University of Technology, Czech Republic; Sanjeev Khudanpur, Center for Language and Speech Processing, United States

Long-span language models that capture syntax and semantics are seldom used in the first pass of large vocabulary continuous speech recognition systems due to the prohibitive search-space of sentence-hypotheses. Instead, an N-best list of hypotheses is created using tractable n-gram models, and rescored using the long-span models. It is shown in this paper that computationally tractable variational approximations of the long-span models are a better choice than standard n-gram models for first pass decoding. They not only result in a better first pass output, but also produce a lattice with a lower oracle word error rate, and rescoring the N-best list from such lattices with the long-span models requires a smaller N to attain the same accuracy. Empirical results on the WSJ, MIT Lectures, NIST 2007 Meeting Recognition and NIST 2001 Conversational Telephone Recognition data sets are presented to support these claims.


  Přepis řeči

|

  Slajdy

Zvětšit slajd | Zobrazit všechny slajdy

0:00:16

  1. slajd

0:02:18

  2. slajd

0:03:11

  3. slajd

0:03:37

  4. slajd

0:05:26

  5. slajd

0:07:40

  6. slajd

0:08:34

  7. slajd

0:09:14

  8. slajd

0:12:36

  9. slajd

0:15:47

 10. slajd

0:18:19

     2. slajd

0:19:18

 11. slajd

0:21:39

     2. slajd

  Komentáře

Please sign in to post your comment!

  Informace o přednášce

Nahráno: 2011-05-25 17:35 - 17:55, Club H
Přidáno: 9. 6. 2011 09:38
Počet zhlédnutí: 41
Rozlišení videa: 1024x576 px, 512x288 px
Délka videa: 0:25:02
Audio stopa: MP3 [8.56 MB], 0:25:02