首页> 外文会议>International conference on text, speech, and dialogue >Reading Comprehension in Czech via Machine Translation and Cross-Lingual Transfer
【24h】

Reading Comprehension in Czech via Machine Translation and Cross-Lingual Transfer

机译:机器翻译和跨语言翻译在捷克语中的阅读理解能力

获取原文

摘要

Reading comprehension is a well studied task, with huge training datasets in English. This work focuses on building reading comprehension systems for Czech, without requiring any manually annotated Czech training data. We then trained and evaluated several BERT and XLM-RoBERTa baseline models. However, our main focus lies in cross-lingual transfer models. We report that a XLM-RoBERTa model trained on English data and evaluated on Czech achieves very competitive performance, only approximately 2% points worse than a model trained on the translated Czech data. This result is extremely good, considering the fact that the model has not seen any Czech data during training. The cross-lingual transfer approach is very flexible and provides a reading comprehension in any language, for which we have enough monolingual raw texts.
机译:阅读理解是一项经过充分研究的任务,拥有大量的英语培训数据集。这项工作着重于为捷克语构建阅读理解系统,而不需要任何人工注释的捷克语培训数据。然后,我们训练并评估了一些BERT和XLM-RoBERTa基线模型。但是,我们的主要重点在于跨语言迁移模型。我们报告说,在英语数据上训练并在捷克语上进行评估的XLM-RoBERTa模型取得了非常好的竞争性能,仅比在翻译后的捷克数据上训练的模型差了约2%。考虑到该模型在训练过程中没有看到任何捷克数据,因此这个结果非常好。跨语言迁移方法非常灵活,可以提供任何语言的阅读理解,为此我们有足够的单语原始文本。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号