首页> 美国政府科技报告 >Tree-Based Hierarchical Reinforcement Learning
【24h】

Tree-Based Hierarchical Reinforcement Learning

机译:基于树的分层强化学习

获取原文

摘要

In this thesis, the author investigates methods for speeding up automatic control algorithms. Specifically, he provides new abstraction techniques for Reinforcement Learning and Semi-Markov Decision Processes (SMDPs). He also introduces the use of policies as temporally abstract actions. This is different from previous definitions of temporally abstract actions as he does not have termination criteria. He provides an approach for processing previously solved problems to extract these policies. He also contributes a method for using supplied or extracted policies to guide and speed up the solving of new problems. He treats extracting policies as a supervised learning task and introduces the Lumberjack algorithm, which extracts repeated sub- structure within a decision tree. He then introduces the TTree algorithm, which combines state and temporal abstraction to increase problem solving speed on new problems. TTree solves SMDPs by using both user- and machine-supplied policies as temporally abstract actions while generating its own tree-based abstract state representation. By combining state and temporal abstraction in this way, TTree is the only known SMDP algorithm that is able to ignore irrelevant or harmful subregions within a supplied abstract action while still making use of other parts of the abstract action.

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号