首页> 外文会议>International Conference on Computational Linguistics >Exploring Question-Specific Rewards for Generating Deep Questions
【24h】

Exploring Question-Specific Rewards for Generating Deep Questions

机译:探索特定问题的奖励,以产生深层问题

获取原文

摘要

Recent question generation (QG) approaches often utilize the sequence-to-sequence framework (Seq2Seq) to optimize the log-likelihood of ground-truth questions using teacher forcing. However, this training objective is inconsistent with actual question quality, which is often reflected by certain global properties such as whether the question can be answered by the document. As such, we directly optimize for QG-specific objectives via reinforcement learning to improve question quality. We design three different rewards that target to improve the fluency, relevance, and answerability of generated questions. We conduct both automatic and human evaluations in addition to a thorough analysis to explore the effect of each QG-specific reward. We find that optimizing question-specific rewards generally leads to better performance in automatic evaluation metrics. However, only the rewards that correlate well with human judgement (e.g., relevance) lead to real improvement in question quality. Optimizing for the others, especially answerability, introduces incorrect bias to the model, resulting in poor question quality.
机译:最近的问题生成(QG)方法通常利用序列到序列框架(SEQ2Seq)来优化使用教师强制性的地面真实问题的日志似然性。然而,这种培训目标与实际问题质量不一致,这通常由某些全局属性反映,例如该问题是否可以由文档回答。因此,我们通过加强学习直接优化QG特定目标,以改善质量。我们设计了三种不同的奖励,该奖励来改善产生的问题的流畅性,相关性和可应答性。除了彻底的分析外,我们还进行自动和人类评估,以探索每个QG特定奖励的效果。我们发现优化质疑奖励通常会导致自动评估指标中的更好性能。然而,只有与人类判断(例如,相关性)相关的奖励导致了质量的真正改善。优化其他,特别是可应答性,将错误的偏差引入模型,导致质量不佳。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号