首页> 外文OA文献 >An integrated approach of learning, planning, and execution
【2h】

An integrated approach of learning, planning, and execution

机译:学习,计划和执行的综合方法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Agents (hardware or software) that act autonomously in an environment have to be able to integrate three basic behaviors: planning, execution, and learning. This integration is mandatory when the agent has no knowledge about how its actions can affect the environment, how the environment reacts to its actions, or, when the agent does not receive as an explicit input, the goals it must achieve. Without an a priori theory, autonomous agents should be able to self-propose goals, set-up plans for achieving the goals according to previously learned models of the agent and the environment, and learn those models from past experiences of successful and failed executions of plans. Planning involves selecting a goal to reach and computing a set of actions that will allow the autonomous agent to achieve the goal. Execution deals with the interaction with the environment by application of planned actions, observation of resulting perceptions, and control of successful achievement of the goals. Learning is needed to predict the reactions of the environment to the agent actions, thus guiding the agent to achieve its goals more efficiently. In this context, most of the learning systems applied to problem solving have been used to learn control knowledge for guiding the search for a plan, but few systems have focused on the acquisition of planning operator descriptions. As an example, currently, one of the most used techniques for the integration of (a way of) planning, execution, and learning is reinforcement learning. However, they usually do not consider the representation of action descriptions, so they cannot reason in terms of goals and ways of achieving those goals. In this paper, we present an integrated architecture, lope, that learns operator definitions, plans using those operators, and executes the plans for modifying the acquired operators. The resulting system is domain-independent, and we have performed experiments in a robotic framework. The results clearly show that the integrated planning, learning, and executing system outperforms the basic planner in that domain.
机译:在环境中自主运行的代理(硬件或软件)必须能够集成三个基本行为:计划,执行和学习。当代理不了解其行为如何影响环境,环境如何对其行为做出反应,或者当代理没有收到明确输入时,必须实现目标时,此集成是必需的。在没有先验理论的情况下,自治主体应该能够提出自己的目标,根据先前了解的主体和环境模型制定实现目标的计划,并从成功执行和失败执行的过去经验中学习这些模型。计划。计划涉及选择要达到的目标并计算一组将使自治代理实现目标的操作。执行通过应用计划的行动,观察所得的看法以及控制目标的成功实现来处理与环境的相互作用。需要学习以预测环境对代理行为的反应,从而指导代理更有效地实现其目标。在这种情况下,大多数用于解决问题的学习系统已用于学习控制知识以指导计划的搜索,但是很少有系统专注于计划操作员描述的获取。作为示例,当前,用于整合(一种方式)计划,执行和学习的最常用技术之一是强化学习。但是,他们通常不考虑动作描述的表示形式,因此他们无法根据目标和实现这些目标的方式进行推理。在本文中,我们提出了一个集成的架构lope,该架构可学习操作员定义,使用这些操作员进行计划并执行用于修改所获取操作员的计划。最终的系统与领域无关,我们已经在机器人框架中进行了实验。结果清楚地表明,在该领域,集成计划,学习和执行系统的性能优于基本计划器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号