首页> 外文会议>International Conference on Language Resources and Evaluation >JASS: Japanese-specific Sequence to Sequence Pre-training for Neural Machine Translation
【24h】

JASS: Japanese-specific Sequence to Sequence Pre-training for Neural Machine Translation

机译:JASS:日本特定序列,用于序列训练神经机翻译

获取原文

摘要

Neural machine translation (NMT) needs large parallel corpora for state-of-the-art translation quality. Low-resource NMT is typically addressed by transfer learning which leverages large monolingual or parallel corpora for pre-training. Monolingual pre-training approaches such as MASS (MAsked Sequence to Sequence) are extremely effective in boosting NMT quality for languages with small parallel corpora. However, they do not account for linguistic information obtained using syntactic analyzers which is known to be invaluable for several Natural Language Processing (NLP) tasks. To this end, we propose JASS, Japanese-specific Sequence to Sequence, as a novel pre-training alternative to MASS for NMT involving Japanese as the source or target language. JASS is joint BMASS (Bunsetsu MASS) and BRSS (Bunsetsu Reordering Sequence to Sequence) pre-training which focuses on Japanese linguistic units called bunsetsus. In our experiments on ASPEC Japanese-English and News Commentary Japanese-Russian translation we show that JASS can give results that are competitive with if not better than those given by MASS. Furthermore, we show for the first time that joint MASS and JASS pre-training gives results that significantly surpass the individual methods indicating their complementary nature. We will release our code, pre-trained models and bunsetsu annotated data as resources for researchers to use in their own NLP tasks.
机译:神经机翻译(NMT)需要大的平行语料库以获得最先进的翻译质量。低资源NMT通常通过转移学习来解决,从而利用大型单机或平行语料库进行预培训。单体式预训练方法,如质量(屏蔽序列)非常有效地提高与小并行语言的语言的NMT质量。但是,它们不考虑使用句法分析仪获得的语言信息,这些信息被称为几种自然语言处理(NLP)任务非常有价值。为此,我们向序列提出Jass,日本特定的序列,作为涉及日语作为源语言或目标语言的NMT的新型预训练替代品。 Jass是联合BMASS(Bunsetsu Mass)和BRSS(Bunsetsu重新排序序列)预先培训,其侧重于叫做Bunsetsus的日本语言单位。在我们的Aspec日语和新闻注释日语 - 俄语翻译中,我们表明JASS可以给出与群众给出的那些竞争的结果。此外,我们首次展示联合群众和JASS预训练提供了显着超越各个方法的结果,这表明其互补性。我们将发布我们的代码,预先训练的模型和Bunsetsu注释数据作为研究人员以自己的NLP任务使用的资源。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号