首页> 外文会议>International Speech Communication Association >Topic Dependent Language Model based on Topic Voting on Noun History
【24h】

Topic Dependent Language Model based on Topic Voting on Noun History

机译:基于主题投票的主题依赖语言模型在名词历史上

获取原文

摘要

Language models (LMs) are important in automatic speech recognition systems. In this paper, we propose a new approach to a topic dependent LM, where the topic is decided in an un-supervised manner. Latent Semantic Analysis (LSA) is employed to reveal hidden (latent) relations among nouns in the context words. To decide the topic of an event, a fixed size word history sequence (window) is observed, and voting is then carried out based on noun class occurrences weighted by a confidence measure. Experiments on the Wall Street Journal corpus and Mainichi Shimbun (Japanese newspaper) corpus show that our proposed method gives better perplexity than the comparative baselines, including a word-based/class-based n-gram LM, their interpolated LM, a cache-based LM, and the Latent Dirichlet Allocation (LDA)-based topic dependent LM.
机译:语言模型(LMS)在自动语音识别系统中很重要。在本文中,我们提出了一种对依赖LM主题的新方法,其中主题以不受监督的方式决定。潜在语义分析(LSA)被用来揭示语境词中名词之间的隐藏(潜在)关系。为了确定事件的主题,观察到固定大小的字历史序列(窗口),然后基于由置信度量加权的名词类出现来执行投票。 Wall Street Journal Corpus和Mainichi Shimbun(日本报纸)语料库的实验表明,我们的提出方法比比较基线提供更好的困惑,包括基于词的/基于类的N-GRAM LM,其内插LM,基于缓存的缓存LM,以及潜在的Dirichlet分配(LDA)基于主题依赖于LM。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号