首页> 外文学位 >A reinforcement learning technique for enhancing human behavior models in a context-based architecture.
【24h】

A reinforcement learning technique for enhancing human behavior models in a context-based architecture.

机译:一种增强学习技术,用于在基于上下文的体系结构中增强人类行为模型。

获取原文
获取原文并翻译 | 示例

摘要

A reinforcement-learning technique for enhancing human behavior models in a context-based learning architecture is presented. Prior to the introduction of this technique, human models built and developed in a Context-Based reasoning framework lacked learning capabilities. As such, their performance and quality of behavior was always limited by what the subject matter expert whose knowledge is modeled was able to articulate or demonstrate. Results from experiments performed show that subject matter experts are prone to making errors and at times they lack information on situations that are inherently necessary for the human models to behave appropriately and optimally in those situations. The benefits of the technique presented is two fold; (1) It shows how human models built in a context-based framework can be modified to correctly reflect the knowledge learnt in a simulator; and (2) It presents a way for subject matter experts to verify and validate the knowledge they share. The results obtained from this research show that behavior models built in a context-based framework can be enhanced by learning and reflecting the constraints in the environment. From the results obtained, it was shown that after the models are enhanced, the agents performed better based on the metrics evaluated. Furthermore, after learning, the agent was shown to recognize unknown situations and behave appropriately in previously unknown situations. The overall performance and quality of behavior of the agent improved significantly. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)
机译:提出了一种增强学习技术,用于在基于上下文的学习体系结构中增强人类行为模型。在引入该技术之前,在基于上下文的推理框架中构建和开发的人体模型缺乏学习能力。因此,他们的行为和行为质量始终受到建模知识的主题专家能够表达或证明的限制。进行的实验结果表明,主题专家容易犯错误,有时他们缺少有关人体模型固有的必要信息,这些条件对于人体模型在这些情况下正确和最佳地运行至关重要。提出的技术的好处有两个方面。 (1)显示了如何修改基于上下文的框架中构建的人体模型,以正确反映在模拟器中学习到的知识; (2)它为主题专家提供了一种验证和验证他们共享的知识的方法。从这项研究中获得的结果表明,可以通过学习并反映环境中的约束来增强在基于上下文的框架中构建的行为模型。从获得的结果可以看出,在对模型进行增强之后,基于评估的指标,代理的表现更好。此外,在学习之后,该代理被证明可以识别未知情况并在以前未知的情况下表现适当。代理的整体性能和行为质量显着提高。 (仅可从麻省理工学院图书馆14-0551室,剑桥,马萨诸塞州02139-4307;电话617-253-5668;传真617-253-1690获得副本。)

著录项

  • 作者

    Aihe, David O. I.;

  • 作者单位

    University of Central Florida.;

  • 授予单位 University of Central Florida.;
  • 学科 Artificial Intelligence.;Computer Science.
  • 学位 Ph.D.
  • 年度 2008
  • 页码 308 p.
  • 总页数 308
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 人工智能理论;自动化技术、计算机技术;
  • 关键词

  • 入库时间 2022-08-17 11:38:46

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号