首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >COMPARING RNNS AND LOG-LINEAR INTERPOLATION OF IMPROVED SKIP-MODEL ON FOUR BABEL LANGUAGES: CANTONESE, PASHTO, TAGALOG, TURKISH
【24h】

COMPARING RNNS AND LOG-LINEAR INTERPOLATION OF IMPROVED SKIP-MODEL ON FOUR BABEL LANGUAGES: CANTONESE, PASHTO, TAGALOG, TURKISH

机译:比较RNN和Log-Log-Log-Log-Log-Log-Log-Linear插值在四种Babel语言上的改进跳过模型:粤语,普什图,Tagalog,土耳其语

获取原文

摘要

Recurrent neural networks (RNNs) are a very recent technique to model long range dependencies in natural languages. They have clearly outperformed trigrams and other more advanced language modeling techniques by using non-linearly modeling long range dependencies. An alternative is to use log-linear interpolation of skip models (i.e. skip bigrams and skip trigrams). The method as such has been published earlier. In this paper we investigate the impact of different smoothing techniques on the skip models as a measure of their overall performance. One option is to use automatically trained distance clusters (both hard and soft) to increase robustness and to combat sparseness in the skip model We also investigate alternative smoothing techniques on word level. For skip bigrams when skipping a small number of words Kneser-Ney smoothing (KN) is advantageous. For a larger number of words being skipped Dirichlet smoothing performs better. In order to exploit the advantages of both KN and Dirichlet smoothing we propose a new unified smoothing technique. Experiments are performed on four Babel languages: Cantonese, Pashto, Tagalog and Turkish. RNNs and log-linearly interpolated skip models are on par if the skip models are trained with standard smoothing techniques. Using the improved smoothing of the skip models along with distance clusters, we can clearly outperform RNNs by about 8-11 % in perplexity across all four languages.
机译:经常性的神经网络(RNNS)是一种最近的技术,用于模拟自然语言的长距离依赖性。它们通过使用非线性建模的长距离依赖性,它们显然优先表现出了三进体和其他更先进的语言建模技术。替代方案是使用跳过模型的日志线性插值(即跳过Bigrams和Skip Trigrams)。此类方法之前发布。在本文中,我们调查了不同平滑技术对跳过模型的影响,作为其整体性能的衡量标准。一种选择是使用自动训练的距离簇(既硬和软),以增加稳健性并在跳过模型中打击稀疏,我们还研究了单词级别的替代平滑技术。对于跳过少量单词的跳过跳过巨大的单词,Ney平滑(KN)是有利的。对于跳过更大数量的单词,Dirichlet Smoothing执行更好。为了利用KN和Dirichlet平滑的优点,我们提出了一种新的统一平滑技术。实验是对四个Babel语言进行的:粤语,普什图,Tagalog和土耳其语。如果跳过型号培训,则RNN和Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Log-Interpatal ock型号具有标准平滑技术。使用跳过模型的改进平滑以及距离集群,我们可以在所有四种语言中显然占据8-11%的困惑。

著录项

相似文献

  • 外文文献
  • 中文文献
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号