首页> 外文会议>Workshop on technologies for MT of low resource languages >System description of Supervised and Unsupervised Neural Machine Translation approaches from 'NL Processing' team at DeepHack.Babel task
【24h】

System description of Supervised and Unsupervised Neural Machine Translation approaches from 'NL Processing' team at DeepHack.Babel task

机译:DeepHack的'NL Processing'团队的有监督和无监督神经机器翻译方法的系统描述

获取原文
获取原文并翻译 | 示例

摘要

A comparison ' of supervised and unsupervised Neural Machine Translation (NMT) models was done for the corpora provided by the DeepHack.Babel competition. It is shown that for even small parallel corpus, fully supervised NMT gives better results than fully unsupervised for the case of constrained domain of the corpus. We have also implemented a fully unsupervised and a semi-supervised NMT models which have not given positive results compared to fully supervised models. A blind set-up is described where participants know at no point what language pair is used for translation, so no extra data could be integrated in pre-submission phase or during training. Finally, future competition organizers should find ways to protect their competition set-ups against various attacks in order to prevent from revealing of language pairs. We have reported two possible types of attacks on the blind set-up.
机译:对DeepHack.Babel竞赛提供的语料库进行了有监督和无监督神经机器翻译(NMT)模型的比较。结果表明,即使是很小的并行语料库,对于受约束的语料域,完全监督的NMT也会比完全不受监督的NMT产生更好的结果。我们还实现了完全无监督的和半监督的NMT模型,与完全监督的模型相比,这些模型没有给出积极的结果。描述了一种盲目设置,参与者不知道该使用哪种语言对进行翻译,因此在提交前阶段或培训期间不会集成任何额外的数据。最后,未来的比赛组织者应该找到保护他们的比赛设置不受各种攻击的方法,以防止语言对的泄露。我们已经报告了针对盲目设置的两种可能的攻击类型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号