首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >The Actor-Dueling-Critic Method for Reinforcement Learning
【2h】

The Actor-Dueling-Critic Method for Reinforcement Learning

机译:强化学习的演员-决斗批评方法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Model-free reinforcement learning is a powerful and efficient machine-learning paradigm which has been generally used in the robotic control domain. In the reinforcement learning setting, the value function method learns policies by maximizing the state-action value (Q value), but it suffers from inaccurate Q estimation and results in poor performance in a stochastic environment. To mitigate this issue, we present an approach based on the actor-critic framework, and in the critic branch we modify the manner of estimating Q-value by introducing the advantage function, such as dueling network, which can estimate the action-advantage value. The action-advantage value is independent of state and environment noise, we use it as a fine-tuning factor to the estimated Q value. We refer to this approach as the actor-dueling-critic (ADC) network since the frame is inspired by the dueling network. Furthermore, we redesign the dueling network part in the critic branch to make it adapt to the continuous action space. The method was tested on gym classic control environments and an obstacle avoidance environment, and we design a noise environment to test the training stability. The results indicate the ADC approach is more stable and converges faster than the DDPG method in noise environments.
机译:无模型强化学习是一种强大而高效的机器学习范例,已广泛用于机器人控制领域。在强化学习设置中,价值函数方法通过最大化状态作用值(Q值)来学习策略,但是它会受到Q估计不准确的困扰,并且在随机环境中会导致性能下降。为了缓解这个问题,我们提出了一种基于参与者批评框架的方法,在批评者分支中,我们通过引入诸如对决网络之类的优势函数来修改估算Q值的方式,该函数可以估算出行动优势值。 。行动优势值与状态噪声和环境噪声无关,我们将其用作对估算Q值的微调因子。由于帧是由决斗网络启发的,因此我们将这种方法称为“演员-决斗批评家(ADC)”网络。此外,我们重新设计了注释器分支中的决斗网络部分,以使其适应连续动作空间。在健身房经典控制环境和避障环境下对该方法进行了测试,并设计了噪声环境以测试训练的稳定性。结果表明,在噪声环境中,ADC方法比DDPG方法更稳定并且收敛速度更快。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号