首页> 外文会议>First workshop on natural machine translation 2017 >An Empirical Study of Mini-Batch Creation Strategies for Neural Machine Translation
【24h】

An Empirical Study of Mini-Batch Creation Strategies for Neural Machine Translation

机译:神经机器翻译的小批量创建策略的实证研究

获取原文
获取原文并翻译 | 示例

摘要

Training of neural machine translation (NMT) models usually uses mini-batches for efficiency purposes. During the mini-batched training process, it is necessary to pad shorter sentences in a mini-batch to be equal in length to the longest sentence therein for efficient computation. Previous work has noted that sorting the corpus based on the sentence length before making mini-batches reduces the amount of padding and increases the processing speed. However, despite the fact that mini-batch creation is an essential step in NMT training, widely used NMT toolkits implement disparate strategies for doing so, which have not been empirically validated or compared. This work investigates mini-batch creation strategies with experiments over two different datasets. Our results suggest that the choice of a mini-batch creation strategy has a large effect on NMT training and some length-based sorting strategies do not always work well compared with simple shuffling.
机译:训练神经机器翻译(NMT)模型通常使用微型批次以提高效率。在小批量训练过程中,为了有效计算,有必要在小批量中填充较短的句子,使其长度等于其中的最长句子。先前的工作已经指出,在进行小批量处理之前,根据句子长度对语料库进行排序可以减少填充量,并提高处理速度。但是,尽管创建小批量生产是NMT培训中必不可少的步骤,但广泛使用的NMT工具箱却采用了不同的策略来进行此操作,但尚未经过经验验证或比较。这项工作通过对两个不同数据集的实验研究了小批量创建策略。我们的结果表明,选择小批量创建策略对NMT培训有很大影响,并且与简单混洗相比,基于长度的排序策略并不总是很好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号