首页> 外文会议>Annual Meeting of the Association for Computational Linguistics;International Joint Conference on natural Language Processing >Learning from Perturbations: Diverse and Informative Dialogue Generation with Inverse Adversarial Training
【24h】

Learning from Perturbations: Diverse and Informative Dialogue Generation with Inverse Adversarial Training

机译:从扰动中学习:多样化和信息性的对话产生,具有反对反对派培训

获取原文

摘要

In this paper, we propose Inverse Adversarial Training (IAT) algorithm for training neural dialogue systems to avoid generic responses and model dialogue history better. In contrast to standard adversarial training algorithms, IAT encourages the model to be sensitive to the perturbation in the dialogue history and therefore learning from perturbations. By giving higher rewards for responses whose output probability reduces more significantly when dialogue history is perturbed, the model is encouraged to generate more diverse and consistent responses. By penalizing the model when generating the same response given perturbed dialogue history, the model is forced to better capture dialogue history and generate more informative responses. Experimental results on two benchmark datasets show that our approach can better model dialogue history and generate more diverse and consistent responses. In addition, we point out a problem of the widely used maximum mutual information (MMI) based methods for improving the diversity of dialogue response generation models and demonstrate it empirically.
机译:在本文中,我们提出了对训练神经对话系统的逆势抗逆性训练(IAT)算法,以避免更好的响应和模型对话史。与标准对抗训练算法相比,IAT鼓励模型对对话历史中的扰动敏感,从而从扰动中学习。通过给予更高的奖励来响应,当对话历史被扰乱时,其输出概率更加明显减少,鼓励模型产生更多样化和一致的反应。通过在给予扰动对话历史记录的同一响应时惩罚模型,该模型被迫更好地捕获对话历史并生成更具信息性的响应。两个基准数据集的实验结果表明,我们的方法可以更好地模型对话历史,并产生更多样化和一致的反应。此外,我们指出了基于广泛使用的基于最大互信息(MMI)的问题,以改善对话响应生成模型的多样性并经验证明它。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号