首页> 外文会议>Workshop on neural generation and translation >Naver Labs Europe's Systems for the Document-Level Generation and Translation Task at WNGT 2019
【24h】

Naver Labs Europe's Systems for the Document-Level Generation and Translation Task at WNGT 2019

机译:Naver Labs Europe在WNGT 2019上的文档级生成和翻译任务系统

获取原文

摘要

Recently, neural models led to significant improvements in both machine translation (MT) and natural language generation tasks (NLG). However, generation of long descriptive summaries conditioned on structured data remains an open challenge. Likewise, MT that goes beyond sentence-level context is still an open issue (e.g., document-level MT or MT with metadata). To address these challenges, we propose to leverage data from both tasks and do transfer learning between MT, NLG, and MT with source-side metadata (MT+NLG). First, we train document-based MT systems with large amounts of parallel data. Then, we adapt these models to pure NLG and MT+NLG tasks by fine-tuning with smaller amounts of domain-specific data. This end-to-end NLG approach, without data selection and planning, outperforms the previous state of the art on the Rotowire NLG task. We participated to the "Document Generation and Translation" task at WNGT 2019, and ranked first in all tracks.
机译:最近,神经模型导致机器翻译(MT)和自然语言生成任务(NLG)的显着改进。但是,以结构化数据为条件的长描述性摘要的生成仍然是一个开放的挑战。同样,超出句子级上下文的MT仍然是一个未解决的问题(例如,文档级MT或带有元数据的MT)。为了应对这些挑战,我们建议利用两个任务的数据,并在MT,NLG和带有源端元数据(MT + NLG)的MT之间进行转移学习。首先,我们训练具有大量并行数据的基于文档的MT系统。然后,我们通过对少量特定于域的数据进行微调,使这些模型适用于纯NLG和MT + NLG任务。这种无需数据选择和计划的端到端NLG方法,优于Rotowire NLG任务的现有技术水平。我们参加了WNGT 2019的“文档生成和翻译”任务,并在所有曲目中排名第一。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号