首页> 外文学位 >Hierarchical reinforcement learning with function approximation for adaptive control.
【24h】

Hierarchical reinforcement learning with function approximation for adaptive control.

机译:具有自适应控制功能逼近的分层强化学习。

获取原文
获取原文并翻译 | 示例

摘要

This dissertation investigates the incorporation of function approximation and hierarchy into reinforcement learning for use in an adaptive control setting through empirical studies.; Reinforcement learning is an artificial intelligence technique whereby an agent discovers which actions lead to optimal task performance through interaction with its environment. Although reinforcement learning is usually employed to find optimal problem solutions in unchanging environments, a reinforcement learning agent can be modified to continually explore and adapt in a dynamic environment, carrying out a form of direct adaptive control. In the adaptive control setting, the reinforcement learning agent must be able to learn and adapt quickly enough to compensate for the dynamics of the environment. Since reinforcement learning is known to converge slowly to optimality in stationary examined as a means to accelerate reinforcement learning. Various levels of agents through the use of function approximation and hierarchical task decomposition. The effectiveness of this approach is tested in simulations of representative reinforcement learning tasks. The comparison of the learning and adaptation provides insight into the suitability of these techniques to accelerate learning and adaptation.; the reinforcement learning agent uses function approximation to store its learned information. The function approximation method chosen provides local generalization, which provides for a controlled diffusion of information throughout the task space. As a consequence, the experiments conducted with function by the amount of information diffusion, can accelerate learning in tasks where similar states call for similar actions.; Hierarchical task decomposition provides a means of representing a task as a set of related subtasks, which introduces modularity into the task's representation not possible in a monolithic representation. One effect of the hierarchy's modularity is to contain certain environment changes within the smaller space of a subtask. Therefore, the experiments comparing hierarchical and monolithic representations of a task demonstrate that the hierarchical representation can accelerate adaptation in response to certain isolated environment changes.
机译:本文通过实证研究,研究了将函数逼近和层次结构并入强化学习中,以用于自适应控制环境。强化学习是一种人工智能技术,通过该技术,代理可以通过与环境的交互来发现哪些操作导致最佳任务性能。尽管通常采用强化学习在不断变化的环境中找到最佳问题解决方案,但是可以对强化学习代理进行修改以在动态环境中不断探索和适应,从而实现直接自适应控制的形式。在自适应控制设置中,强化学习代理必须能够足够快地学习和适应,以补偿环境的动态变化。由于已知强化学习会慢慢收敛到最优状态,因此可以加快强化学习的速度。通过使用功能逼近和分层任务分解来实现各个级别的代理。在代表性的强化学习任务的模拟中测试了这种方法的有效性。学习和适应的比较提供了对这些技术是否适用于加速学习和适应的见解。强化学习代理使用函数逼近来存储其学习的信息。所选的函数逼近方法可提供局部概括,从而可在整个任务空间中实现信息的受控扩散。结果,通过信息扩散的数量进行功能的实验可以加速任务在相似状态要求相似动作的学习。分层任务分解提供了一种将任务表示为一组相关子任务的方法,这将模块性引入了单块表示法中不可能实现的任务表示。层次结构模块化的作用是将某些环境更改包含在子任务的较小空间内。因此,比较任务的分层表示和整体表示的实验表明,分层表示可以响应某些孤立的环境变化来加速适应。

著录项

  • 作者

    Skelly, Margaret Mary.;

  • 作者单位

    Case Western Reserve University.;

  • 授予单位 Case Western Reserve University.;
  • 学科 Engineering Electronics and Electrical.; Artificial Intelligence.
  • 学位 Ph.D.
  • 年度 2004
  • 页码 261 p.
  • 总页数 261
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 无线电电子学、电信技术;人工智能理论;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号