首页> 外文期刊>Machine translation >Optimizing segmentation granularity for neural machine translation
【24h】

Optimizing segmentation granularity for neural machine translation

机译:神经电机翻译优化分割粒度

获取原文
获取原文并翻译 | 示例
           

摘要

In neural machine translation (NMT), it has become standard to translate using subword units to allow for an open vocabulary and improve accuracy on infrequent words. Byte-pair encoding (BPE) and its variants are the predominant approach to generating these subwords, as they are unsupervised, resource-free, and empirically effective. However, the granularity of these subword units is a hyperparameter to be tuned for each language and task, using methods such as grid search. Tuning may be done inexhaustively or skipped entirely due to resource constraints, leading to sub-optimal performance. In this paper, we propose a method to automatically tune this parameter using only one training pass. We incrementally introduce new BPE vocabulary online based on the held-out validation loss, beginning with smaller, general subwords and adding larger, more specific units over the course of training. Our method matches the results found with grid search, optimizing segmentation granularity while significantly reducing overall training time. We also show benefits in training efficiency and performance improvements for rare words due to the way embeddings for larger units are incrementally constructed by combining those from smaller units.
机译:在神经机翻译(NMT)中,它已成为标准,可以使用子字单元进行翻译,以允许开放的词汇,并提高不频繁单词的准确性。字节对编码(BPE)及其变体是生成这些子字的主要方法,因为它们是无监督,无资资源和凭证有效的。但是,这些子字单元的粒度是使用诸如网格搜索的方法进行调整的超级参数。由于资源限制,可以完全完成调整或完全跳过,从而导致次优性能。在本文中,我们提出了一种使用只使用一个训练通过自动调整此参数的方法。我们逐步在线在线在线在线在线在线在线,从较小的常规次字开始,并在培训过程中添加更大的更具体的单位。我们的方法与网格搜索的结果匹配,优化分段粒度,同时显着降低整体培训时间。我们还显示培训效率和稀有单词的性能改善由于较大单位的嵌入方式通过与较小单位组合来逐步构建。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号