首页> 外文会议>Conference on the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >Learning When Not to Answer: A Ternary Reward Structure for Reinforcement Learning based Question Answering
【24h】

Learning When Not to Answer: A Ternary Reward Structure for Reinforcement Learning based Question Answering

机译:学习何时不回答:基于学习的问答式强化三元奖励结构

获取原文

摘要

In this paper, we investigate the challenges of using reinforcement learning agents for question-answering over knowledge graphs for real-world applications. We examine the performance metrics used by state-of-the-art systems and determine that they are inadequate for such settings. More specifically, they do not evaluate the systems correctly for situations when there is no answer available and thus agents optimized for these metrics are poor at modeling confidence. We introduce a simple new performance metric for evaluating question-answering agents that is more representative of practical usage conditions, and optimize for this metric by extending the binary reward structure used in prior work to a ternary reward structure which also rewards an agent for not answering a question rather than giving an incorrect answer. We show that this can drastically improve the precision of answered questions while only not answering a limited number of previously correctly answered questions. Employing a supervised learning strategy using depth-first-search paths to bootstrap the reinforcement learning algorithm further improves performance.
机译:在本文中,我们调查了使用强化学习代理对现实应用中的知识图进行问题解答的挑战。我们检查了最新系统使用的性能指标,并确定它们不足以进行此类设置。更具体地说,在没有可用答案的情况下,他们无法正确评估系统,因此针对这些指标进行优化的代理在建模信心方面很差。我们引入了一个用于评估问答代理的简单新绩效指标,该指标更能代表实际使用条件,并通过将先前工作中使用的二进制奖励结构扩展为三元奖励结构来对该指标进行优化,该三元奖励结构还会奖励未回答的代理商一个问题,而不是给出错误的答案。我们表明,这可以大大提高回答问题的准确性,而仅不回答有限数量的先前正确回答的问题。使用使用深度优先搜索路径的监督学习策略来引导强化学习算法,可以进一步提高性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号