【24h】

Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning

机译:产生良好计划的学习方法:HTN学习与强化学习的整合

获取原文

摘要

We consider how to learn Hierarchical Task Networks (HTNs) for planning problems in which both the quality of solution plans generated by the HTNs and the speed at which those plans are found is important. We describe an integration of HTN Learning with Reinforcement Learning to both learn methods by analyzing semantic annotations on tasks and to produce estimates of the expected values of the learned methods by performing Monte Carlo updates. We performed an experiment in which plan quality was inversely related to plan length. In two planning domains, we evaluated the planning performance of the learned methods in comparison to two state-of-the-art satisficing classical planners, FastForward and SgPlan6, and one optimal planner, HSP_F. The results demonstrate that a greedy HTN planner using the learned methods was able to generate higher quality solutions than SgPlan6 in both domains and FASTFORWARD in one. Our planner, FastForward, and SGPLAN6 ran in similar time, while HSP_F~* was exponentially slower.
机译:我们考虑如何学习用于计划问题的分层任务网络(HTN),在这些问题中,由HTN生成的解决方案计划的质量和找到这些计划的速度都非常重要。我们描述了HTN学习与强化学习的集成,以通过分析任务上的语义注释来学习方法,并通过执行蒙特卡洛更新来生成学习方法的期望值的估计值。我们进行了一项实验,其中计划质量与计划长度成反比。在两个规划领域中,我们与两个最先进的令人满意的经典规划器FastForward和SgPlan6以及一个最佳规划器HSP_F进行了比较,评估了所学方法的规划性能。结果表明,使用学习方法的贪婪的HTN计划程序能够在两个域中都比SgPlan6和FASTFORWARD中的一个生成更高质量的解决方案。我们的计划程序FastForward和SGPLAN6的运行时间相似,而HSP_F〜*的运行速度却成倍下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号