首页> 外文期刊>Mathematical Problems in Engineering: Theory, Methods and Applications >Deep Q-Network with Predictive State Models in Partially Observable Domains
【24h】

Deep Q-Network with Predictive State Models in Partially Observable Domains

机译:具有部分可观察域中的预测状态模型的深Q网络

获取原文
           

摘要

While deep reinforcement learning (DRL) has achieved great success in some large domains, most of the related algorithms assume that the state of the underlying system is fully observable. However, many real-world problems are actually partially observable. For systems with continuous observation, most of the related algorithms, e.g., the deep Q-network (DQN) and deep recurrent Q-network (DRQN), use history observations to represent states; however, they often make computation-expensive and ignore the information of actions. Predictive state representations (PSRs) can offer a powerful framework for modelling partially observable dynamical systems with discrete or continuous state space, which represents the latent state using completely observable actions and observations. In this paper, we present a PSR model-based DQN approach which combines the strengths of the PSR model and DQN planning. We use a recurrent network to establish the recurrent PSR model, which can fully learn dynamics of the partially continuous observable environment. Then, the model is used for the state representation and update of DQN, which makes DQN no longer rely on a fixed number of history observations or recurrent neural network (RNN) to represent states in the case of partially observable environments. The strong performance of the proposed approach is demonstrated on a set of robotic control tasks from OpenAI Gym by comparing with the technique with the memory-based DRQN and the state-of-the-art recurrent predictive state policy (RPSP) networks. Source code is available at https://github.com/RPSR-DQN/paper-code.git.
机译:虽然深度加强学习(DRL)在一些大域中取得了巨大的成功,但大多数相关算法假设底层系统的状态是完全可观察的。然而,许多真实世界问题实际上是部分观察到的。对于具有连续观察的系统,大多数相关算法,例如,深度Q-Network(DQN)和深度复发性Q-Network(DRQN),使用历史观察来表示状态;但是,它们通常会使计算昂贵并忽略动作的信息。预测状态表示(PSR)可以提供一种强大的框架,用于使用离散或连续状态空间建模部分可观察的动态系统,其表示使用完全可观察的动作和观察的潜在状态。在本文中,我们提出了一种基于PSR模型的DQN方法,它结合了PSR模型和DQN规划的优势。我们使用经常性网络来建立复发性PSR模型,可以完全学习部分连续可观察环境的动态。然后,该模型用于DQN的状态表示和更新,这使得DQN不再依赖于固定数量的历史观察或经常性神经网络(RNN)来代表部分可观察环境的情况。通过与基于内存的DRQN的技术与现有的经常性预测状态政策(RPSP)网络的技术相比,在Openai Bumb的一组机器人控制任务中对所提出的方法的强劲表明。源代码可在https://github.com/rpsr-dqn/paper-code.git中获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号