首页> 外文会议> >Memory-guided exploration in reinforcement learning
【24h】

Memory-guided exploration in reinforcement learning

机译:记忆学习中的强化学习探索

获取原文

摘要

We focus on the task transfer in reinforcement learning and specifically in Q-learning. There are three main model free methods for performing task transfer in Q-learning: direct transfer, soft transfer and memory-guided exploration. In direct transfer, the Q-values from a previous task are used to initialize the Q-values of the next task. The soft transfer initializes the Q-values of the new task with a weighted average of the standard initialization value and the Q-values of the previous task. In memory-guided exploration the Q-values of previous tasks are used as a guide in the initial exploration of the agent. The weight that the agent gives to its past experience decreases over time. We explore stability issues related to the off-policy nature of memory-guided exploration and compare memory-guided exploration to soft transfer and direct transfer in three different environments.
机译:我们专注于强化学习中的任务转移,尤其是Q学习中的任务转移。在Q学习中执行任务转移有三种主要的无模型方法:直接转移,软转移和内存引导的探索。在直接传输中,上一个任务的Q值用于初始化下一个任务的Q值。软传输使用标准初始化值和先前任务的Q值的加权平均值来初始化新任务的Q值。在以内存为指导的探索中,先前任务的Q值将用作对代理进行初始探索的指导。代理赋予其过去经验的权重会随着时间的推移而降低。我们探讨了与内存引导的探索的非策略性相关的稳定性问题,并将内存引导的探索与软传输和直接传输在三种不同环境中进行了比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号