首页> 外文会议>Machine translation summit >Improving Neural Machine Translation Using Noisy Parallel Data through Distillation
【24h】

Improving Neural Machine Translation Using Noisy Parallel Data through Distillation

机译:通过蒸馏使用嘈杂的并行数据改​​进神经机翻译

获取原文

摘要

Due to the scarcity of parallel training data for many language pairs, quasi-parallel or comparable training data provides an important alternative resource for training machine translation systems for such language pairs. Since comparable corpora are not of as high quality as manually annotated parallel data, using them for training can have a negative effect on the translation performance of an NMT model. We propose distillation as a remedy to effectively leverage comparable data where the training of a student model on combined clean and comparable data is guided by a teacher model trained on the high-quality, clean data only. Our experiments for Arabic-English, Chinese-English, and German-English translation demonstrate that distillation yields significant improvements compared to off-the-shelf use of comparable data and performs comparable to state-of-the-art methods for noise filtering.
机译:由于许多语言对的并行训练数据的稀缺性,准平行或可比较的训练数据为这种语言对提供了一种重要的替代资源,用于培训机器翻译系统。由于可比的Corpora不像手动注释的并行数据的高品质,因此使用它们进行培训可能对NMT模型的翻译性能产生负面影响。我们提出蒸馏作为一个补救措施,以有效利用可比的数据,其中学生模型的组合清洁和可比数据的培训是由高质量,清洁数据训练的教师模型引导的。我们对阿拉伯语 - 英语,中文和德语 - 英语翻译的实验表明,与特征使用可比数据相比,蒸馏产生了显着的改善,并执行与最先进的噪声滤波方法相当的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号