首页> 外文会议>International Conference on Machine Learning >Mixture Models for Diverse Machine Translation: Tricks of the Trade
【24h】

Mixture Models for Diverse Machine Translation: Tricks of the Trade

机译:不同机器翻译的混合模型:贸易的技巧

获取原文

摘要

Mixture models trained via EM are among the simplest, most widely used and well understood latent variable models in the machine learning literature. Surprisingly, these models have been hardly explored in text generation applications such as machine translation. In principle, they provide a latent variable to control generation and produce a diverse set of hypotheses. In practice, however, mixture models are prone to degeneracies - often only one component gets trained or the latent variable is simply ignored. We find that disabling dropout noise in responsibility computation is critical to successful training. In addition, the design choices of parameterization, prior distribution, hard versus soft EM and online versus offline assignment can dramatically affect model performance. We develop an evaluation protocol to assess both quality and diversity of generations against multiple references, and provide an extensive empirical study of several mixture model variants. Our analysis shows that certain types of mixture models are more robust and offer the best trade-off between translation quality and diversity compared to variational models and diverse decoding approaches.
机译:通过EM培训的混合物模型是机器学习文献中最简单,最广泛使用和最良好的潜在变量模型之一。令人惊讶的是,这些模型在文本生成应用中几乎没有探索,例如机器翻译。原则上,它们提供了控制生成的潜在变量,并产生各种假设。然而,在实践中,混合模型易于退化 - 通常只有一个组件受过训练,或者潜在的变量简单地忽略。我们发现,禁用责任计算的丢失噪声对于成功培训至关重要。此外,参数化的设计选择,先前分布,硬质与软件和在线与离线分配可以显着影响模型性能。我们开发评估议定书,以评估对多种参考的几代质量和多样性,并对几种混合物模型变体提供广泛的实证研究。我们的分析表明,与变分模型和多样化的解码方法相比,某些类型的混合模型更加强劲,并在翻译质量和多样性之间提供最佳权衡。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号