...
首页> 外文期刊>Artificial intelligence >Local and global explanations of agent behavior: Integrating strategy summaries with saliency maps
【24h】

Local and global explanations of agent behavior: Integrating strategy summaries with saliency maps

机译:代理行为的本地和全球解释:将战略摘要与显着性图集成在一起

获取原文
获取原文并翻译 | 示例

摘要

With advances in reinforcement learning (RL), agents are now being developed in high-stakes application domains such as healthcare and transportation. Explaining the behavior of these agents is challenging, as the environments in which they act have large state spaces, and their decision-making can be affected by delayed rewards, making it difficult to analyze their behavior. To address this problem, several approaches have been developed. Some approaches attempt to convey the global behavior of the agent, describing the actions it takes in different states. Other approaches devised local explanations which provide information regarding the agent's decision-making in a particular state. In this paper, we combine global and local explanation methods, and evaluate their joint and separate contributions, providing (to the best of our knowledge) the first user study of combined local and global explanations for RL agents. Specifically, we augment strategy summaries that extract important trajectories of states from simulations of the agent with saliency maps which show what information the agent attends to. Our results show that the choice of what states to include in the summary (global information) strongly affects people's understanding of agents: participants shown summaries that included important states significantly outperformed participants who were presented with agent behavior in a set of world-states that are likely to appear during gameplay. We find mixed results with respect to augmenting demonstrations with saliency maps (local information), as the addition of saliency maps, in the form of raw heat maps, did not significantly improve performance in most cases. However, we do find some evidence that saliency maps can help users better understand what information the agent relies on during its decision-making, suggesting avenues for future work that can further improve explanations of RL agents.
机译:随着钢筋学习(RL)的进步,现在正在高赌注应用领域开发代理,例如医疗保健和运输。解释这些代理的行为是具有挑战性的,因为它们的行为具有大状态空间,而他们的决策可能会受到延迟奖励的影响,使其难以分析其行为。为了解决这个问题,已经开发了几种方法。一些方法尝试传达代理的全球行为,描述它所带来的不同状态。其他方法设计了当地的解释,该解释提供有关代理人在特定状态下的决策的信息。在本文中,我们结合了全球和地方解释方法,评估了他们的联合和单独的贡献,提供了(据我们所知)的第一个用户研究,对RL代理商的合并本地和全球解释。具体而言,我们增强了从具有显着性图的代理模拟提取各种轨迹的策略摘要,这些概率显示了代理商出席的信息。我们的研究结果表明,在摘要(全球信息)中的选择强烈影响人们对代理人的理解:参与者所表明的摘要,其中包括重要国家在一组世界陈述中呈现出具有代理行为的重要态度显着优势。可能会出现在游戏过程中。我们在显着图(本地信息)的增强示范中找到了混合结果,因为在原始热图的形式增加了显着图,在大多数情况下没有显着提高性能。但是,我们确实发现了一些证据表明,显着性图可以帮助用户更好地了解代理在决策期间依赖的信息,建议可以进一步改进对RL代理商的解释的途径。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号