【24h】

Invited Speaker

机译:邀请扬声器

获取原文

摘要

Recently Neural Machine Translation (NMT) systems are reported to outperform other approaches in machine translation. However, NMT systems are known to be computationally expensive both in training and in translation inference - sometimes prohibitively so in the case of very large data sets and large models. Several authors have also charged that NMT systems lack robustness, particularly when input sentences contain rare words. These issues have hindered NMT's use in practical deployments and services, where both accuracy and speed are essential. In this talk, I present GNMT, Google's Neural Machine Translation system, which attempts to address many of these issues. Our model consists of a deep LSTM network with 8 encoder and 8 decoder layers using residual connections as well as attention connections from the decoder network to the encoder. To improve parallelism and therefore decrease training time, our attention mechanism connects the bottom layer of the decoder to the top layer of the encoder. To accelerate the final translation speed, we employ low-precision arithmetic during inference computations. To improve handling of rare words, we divide words into a limited set of common sub-word units ("wordpieces") for both input and output. On the WMT'14 English-to-French and English-to-German benchmarks, GNMT achieves competitive results to state-of-the-art. Using a human side-by-side evaluation on a set of isolated simple sentences, it reduces translation errors by an average of 60phrase-based production system.
机译:据报道,最近神经机翻译(NMT)系统以机器翻译中的其他方法优于其他方法。然而,已知NMT系统在训练和翻译推理中既可以计算昂贵 - 有时在非常大的数据集和大型模型的情况下。若干作者还指控NMT系统缺乏稳健性,特别是当输入句子包含稀有单词时。这些问题阻碍了NMT在实际部署和服务中的使用,其中精度和速度都是必不可少的。在这次谈话中,我展示了GNMT,谷歌的神经机翻译系统,试图解决许多这些问题。我们的模型包括一个带有8个编码器和8个解码器层的深层LSTM网络,以及使用残留连接以及从解码器网络到编码器的注意连接。为了改善平行性并因此降低训练时间,我们的注意机构将解码器的底层连接到编码器的顶层。为了加速最终的转换速度,我们在推理计算期间采用低精度算术。为了改善稀有单词的处理,我们将单词划分为有限的常见子字单元(“WordPieces”),用于输入和输出。在WMT'14英语到法国和英语到德国基准中,GNMT实现了最先进的竞争结果。在一组孤立的简单句子上使用人类并排评估,它通过平均基于60phrase的生产系统来降低转换误差。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号