首页> 外文期刊>Computational intelligence and neuroscience >End-to-End Autonomous Exploration with Deep Reinforcement Learning and Intrinsic Motivation
【24h】

End-to-End Autonomous Exploration with Deep Reinforcement Learning and Intrinsic Motivation

机译:基于深度强化学习和内在动机的端到端自主探索

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Developing artificial intelligence (AI) agents is challenging for efficient exploration in visually rich and complex environments. In this study, we formulate the exploration question as a reinforcement learning problem and rely on intrinsic motivation to guide exploration behavior. Such intrinsic motivation is driven by curiosity and is calculated based on episode memory. To distribute the intrinsic motivation, we use a count-based method and temporal distance to generate it synchronously. We tested our approach in 3D maze-like environments and validated its performance in exploration tasks through extensive experiments. The experimental results show that our agent can learn exploration ability from raw sensory input and accomplish autonomous exploration across different mazes. In addition, the learned policy is not biased by stochastic objects. We also analyze the effects of different training methods and driving forces on exploration policy.
机译:开发人工智能 (AI) 代理对于在视觉丰富和复杂的环境中进行高效探索具有挑战性。在这项研究中,我们将探索问题表述为强化学习问题,并依靠内在动机来指导探索行为。这种内在动机是由好奇心驱动的,是根据情节记忆计算得出的。为了分配内在动机,我们使用基于计数的方法和时间距离来同步生成它。我们在类似3D迷宫的环境中测试了我们的方法,并通过广泛的实验验证了其在探索任务中的性能。实验结果表明,智能体能够从原始感官输入中学习探索能力,实现跨不同迷宫的自主探索。此外,学习到的策略不会受到随机对象的影响。本文还分析了不同训练方法和驱动力对勘探政策的影响。

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号