首页> 外文会议>Workshop on Structured Prediction for NLP >Log-Linear Reformulation of the Noisy Channel Model for Document-Level Neural Machine Translation
【24h】

Log-Linear Reformulation of the Noisy Channel Model for Document-Level Neural Machine Translation

机译:记录文档级神经机翻译噪声频道模型的对数线性重构

获取原文

摘要

We seek to maximally use various data sources, such as parallel and monolingual data, to build an effective and efficient document-level translation system. In particular, we start by considering a noisy channel approach (Yu et al., 2020) that combines a target-to-source translation model and a language model. By applying Bayes' rule strategically, we reformulate this approach as a log-linear combination of translation, sentence-level and document-level language model probabilities. In addition to using static coefficients for each term, this formulation alternatively allows for the learning of dynamic per-token weights to more finely control the impact of the language models. Using both static or dynamic coefficients leads to improvements over a context-agnostic baseline and a context-aware concatenation model.
机译:我们寻求最大限度地使用各种数据源,例如并行和单格式数据,以构建有效和有效的文档级别转换系统。特别是,我们首先考虑嘈杂的频道方法(Yu等,2020),它结合了目标到源转换模型和语言模型。通过策略性地应用贝叶斯规则,我们将这种方法重构为翻译,句子级和文档级语言模型概率的逻辑线性组合。除了每个术语的静态系数之外,该配方允许允许学习动态的每令牌权重,以更细化语言模型的影响。使用静态或动态系数导致改进上下文无关基线和上下文感知的替代模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号