首页> 外文会议>20th European conference on artificial intelligence >A Robust Approach to Addressing Human Adversaries in Security Games
【24h】

A Robust Approach to Addressing Human Adversaries in Security Games

机译:解决安全游戏中人类对手的稳健方法

获取原文
获取原文并翻译 | 示例

摘要

Game-theoretic approaches have been proposed for addressing the complex problem of assigning limited security resources to protect a critical set of targets. However, many of the standard assumptions fail to address human adversaries who security forces will likely face. To address this challenge, previous research has attempted to integrate models of human decision-making into the game-theoretic algorithms for security settings. The current leading approach, based on experimental evaluation, is derived from a well-founded solution concept known as quantal response and is known as BRQR. One critical difficulty with opponent modeling in general is that, in security domains, information about potential adversaries is often sparse or noisy and furthermore, the games themselves are highly complex and large in scale. Thus, we chose to examine a completely new approach to addressing human adversaries that avoids the complex task of modeling human decision-making. We leverage and modify robust optimization techniques to create a new type of optimization where the defender's loss for a potential deviation by the attacker is bounded by the distance of that deviation from the expected-value-maximizing strategy. To demonstrate the advantages of our approach, we introduce a systematic way to generate meaningful reward structures and compare our approach with BRQR in the most comprehensive investigation to date involving 104 security settings where previous work has tested only up to 10 security settings. Our experimental analysis reveals our approach performing as well as or outperforming BRQR in over 90% of the security settings tested and we demonstrate significant runtime benefits. These results are in favor of utilizing an approach based on robust optimization in these complex domains to avoid the difficulties of opponent modeling.
机译:已经提出了博弈论方法来解决分配有限的安全资源以保护一组关键目标的复杂问题。但是,许多标准假设都无法解决安全部队可能面临的人类对手。为了应对这一挑战,先前的研究尝试将人类决策模型集成到用于安全设置的博弈论算法中。当前基于实验评估的领先方法是从众所周知的解决方案概念(称为量化响应)衍生而来的,该解决方案称为BRQR。通常,对手建模的一个关键困难是,在安全领域中,有关潜在对手的信息通常很少或嘈杂,此外,游戏本身非常复杂且规模庞大。因此,我们选择研究一种全新的解决人类对手的方法,该方法避免了建模人类决策的复杂任务。我们利用并修改健壮的优化技术来创建一种新型的优化,其中防御者因攻击者的潜在偏差所遭受的损失受该偏差与期望值最大化策略的距离的限制。为了证明我们方法的优势,我们引入了一种系统化的方法来生成有意义的奖励结构,并在迄今为止涉及104个安全设置的最全面调查中将我们的方法与BRQR进行比较,以前的工作最多只能测试10个安全设置。我们的实验分析表明,在超过90%的测试安全设置中,我们的方法在BRQR方面的表现均达到或优于BRQR,并且在运行时间方面具有明显优势。这些结果有助于在这些复杂域中使用基于鲁棒优化的方法来避免对手建模的困难。

著录项

  • 来源
  • 会议地点 Montpellier(FR)
  • 作者单位

    University of Southern California, Los Angeles, CA 90089;

    University of Southern California, Los Angeles, CA 90089;

    University of Southern California, Los Angeles, CA 90089;

    University of Southern California, Los Angeles, CA 90089;

    University of Southern California, Los Angeles, CA 90089,Bar-Ilan University, Ramat-Gan 52900, Israel and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 20742;

  • 会议组织
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号