首页> 外文会议>AAAI Conference on Artificial Intelligence >Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning
【24h】

Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning

机译:生成良好计划的学习方法:整合HTN学习和加强学习

获取原文

摘要

We consider how to learn Hierarchical Task Networks (HTNs) for planning problems in which both the quality of solution plans generated by the HTNs and the speed at which those plans are found is important. We describe an integration of HTN Learning with Reinforcement Learning to both learn methods by analyzing semantic annotations on tasks and to produce estimates of the expected values of the learned methods by performing Monte Carlo updates. We performed an experiment in which plan quality was inversely related to plan length. In two planning domains, we evaluated the planning performance of the learned methods in comparison to two state-of-the-art satisficing classical planners, FASTFORWARD and SGPLAN6, and one optimal planner, HSP*F. The results demonstrate that a greedy HTN planner using the learned methods was able to generate higher quality solutions than SGPLAN6 in both domains and FASTFORWARD in one. Our planner, FASTFORWARD, and SGPLAN6 ran in similar time, while HSP*F was exponentially slower.
机译:我们考虑如何学习分层任务网络(HTNS),以规划问题的问题,其中HTNS生成的解决方案计划的质量和那些计划的速度很重要。我们通过分析任务的语义注释和通过执行Monte Carlo更新来描述HTN学习与钢筋学习的集成,以通过分析对任务的语义注释,并通过执行Monte Carlo更新来产生所学习方法的预期值的估计。我们进行了一个实验,计划质量与计划长度相反。在两个规划域中,我们评估了学习方法的规划性能与两个最先进的古典规划者,快速和SGPLAN6以及一个最优规划师,HSP * F的态度相比。结果表明,使用学习方法的贪婪HTN规划器能够在两个域中产生比SGPLAN6更高的质量解决方案,并且在一个方面快速。我们的策划者,Fastforward和Sgplan6类似的时间,而HSP * F是指数较慢的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号