首页> 外文OA文献 >Size matters: An empirical study of neural network training for large vocabulary continuous speech recognition
【2h】

Size matters: An empirical study of neural network training for large vocabulary continuous speech recognition

机译:大小很重要:对大词汇量连续语音识别进行神经网络训练的实证研究

摘要

We have trained and tested a number of large neural networks for the purpose of emission probability estimation in large vocabulary continuous speech recognition. In particular, the problem under test is the DARPA Broadcast News task. Our goal here was to determine the relationship between training time, word error rate, size of the training set, and size of the neural network. In all cases, the network architecture was quite simple, comprising a single large hidden layer with an input window consisting of feature vectors from 9 frames around the current time, with a single output for each of 54 phonetic categories. Thus far, simultaneous increases to the size of the training set and the neural network improve performance; in other words, more data helps, as does the training of more parameters. We continue to be surprised that such a simple system works as well as it does for complex tasks. Given a limitation in training time, however, there appears to be an optimal ratio of training patterns to parameters of around 25:1 in these circumstances. Additionally, doubling the training data and system size appears to provide diminishing returns of error rate reduction for the largest systems.
机译:为了训练大词汇量连续语音识别中的发射概率,我们已经训练和测试了许多大型神经网络。特别是,正在测试的问题是DARPA广播新闻任务。我们的目标是确定训练时间,单词错误率,训练集大小和神经网络大小之间的关系。在所有情况下,网络体系结构都非常简单,包括一个大的隐藏层,该隐藏层的输入窗口由当前时间周围9帧的特征向量组成,并为54个语音类别的每一个提供单个输出。到目前为止,同时增加训练集的大小和神经网络可以改善性能。换句话说,更多的数据有帮助,训练更多的参数也有帮助。对于这样一个简单的系统以及完成复杂任务的效果,我们仍然感到惊讶。但是,鉴于培训时间的限制,在这种情况下,培训模式与参数的最佳比例似乎约为25:1。此外,将训练数据和系统大小加倍似乎可以减少大型系统的错误率降低回报。

著录项

  • 作者单位
  • 年度 1999
  • 总页数
  • 原文格式 PDF
  • 正文语种 {"code":"sq","name":"Albanian","id":41}
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号