首页> 外文会议>IEEE Global Conference on Consumer Electronics >Syllable-Level Long Short-Term Memory Recurrent Neural Network-based Language Model for Korean Voice Interface in Intelligent Personal Assistants
【24h】

Syllable-Level Long Short-Term Memory Recurrent Neural Network-based Language Model for Korean Voice Interface in Intelligent Personal Assistants

机译:基于音节水平的长期短期记忆递归神经网络的智能个人助理中韩语语音接口的语言模型

获取原文

摘要

This study proposes a syllable-level long short-term memory (LSTM) recurrent neural network (RNN)-based language model for a Korean voice interface in intelligent personal assistants (IPAs). Most Korean voice interfaces in IPAs use word-level $n$ -gram language models. Such models suffer from the following two problems: 1) the syntax information in a longer word history is limited because of the limitation of $n$ and 2) The out-of-vocabulary (OOV) problem can occur in a word-based vocabulary. To solve the first problem, the proposed model uses an LSTM RNN-based language model because an LSTM RNN provides long-term dependency information. To solve the second problem, the proposed model is trained with a syllable-level text corpus. Korean words comprise syllables, and therefore, OOV words are not presented in a syllable-based lexicon. In experiments, the RNN-based language model and the proposed model achieved perplexity (PPL) of 68.74 and 17.81, respectively.
机译:这项研究为智能个人助理(IPA)中的韩语语音接口提出了基于音节级的长期短期记忆(LSTM)递归神经网络(RNN)的语言模型。 IPA中的大多数韩语语音接口使用单词级 $ n $ -gram语言模型。这样的模型存在以下两个问题:1)由于以下方面的限制,限制了较长单词历史中的语法信息: $ n $ 2)基于单词的词汇中可能会出现词汇不足(OOV)问题。为了解决第一个问题,建议的模型使用基于LSTM RNN的语言模型,因为LSTM RNN提供了长期依​​赖信息。为了解决第二个问题,用音节级文本语料库对提出的模型进行训练。朝鲜语单词包含音节,因此,在基于音节的词典中不会显示OOV单词。在实验中,基于RNN的语言模型和建议的模型分别实现了68.74和17.81的困惑度(PPL)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号