首页> 外文期刊>Internet of Things Journal, IEEE >CDDPG: A Deep-Reinforcement-Learning-Based Approach for Electric Vehicle Charging Control
【24h】

CDDPG: A Deep-Reinforcement-Learning-Based Approach for Electric Vehicle Charging Control

机译:CDDPG:一种基于深加固的电动汽车充电控制方法

获取原文
获取原文并翻译 | 示例
       

摘要

Electric vehicle (EV) has become one of the most critical components in the smart grid with the applications of the Internet-of-Things (IoT) technologies. Real-time charging control is pivotal to ensure the efficient operation of EVs. However, the charging control performance is limited by the uncertainty of the environment. On the other hand, it is challenging to determine a charging control strategy that is able to optimize multiple objectives simultaneously. In this article, we formulate the EV charging control model as a Markov decision process (MDP) by constructing state, action, transition function, and reward. Then, we propose a deep-reinforcement-learning-based approach: charging control deep deterministic policy gradient (CDDPG) to learn the optimal charging control strategy for satisfying the user's requirement of battery energy while minimizing the user's charging expense. We utilize the long short-term memory (LSTM) network that extracts the information of previous energy price to determine the current charging control strategy. Moreover, Gaussian noise is added to the output of the actor network to prevent the agent from sticking into the nonoptimal strategy. In addition, we address the limitation of sparse rewards by using two replay buffers, of which one is used to store the rewards during the charging phase and another is used to store the rewards after charging is completed. The simulation results prove that the CDDPG-based approach outperforms the deep-Q-learning-based approach (DQL) and the deep-deterministic-policy-gradient-based approach (DDPG) in satisfying the user's requirement for the battery energy and reducing the charging cost.
机译:电动车(EV)已成为智能电网中最关键的组件之一,具有互联网的应用程序(物联网)技术。实时充电控制是关键的,以确保EV的有效操作。然而,充电控制性能受到环境不确定性的限制。另一方面,确定能够同时优化多个目标的充电控制策略充满挑战。在本文中,我们通过构建状态,动作,转换功能和奖励将EV充电控制模型作为马尔可夫决策过程(MDP)制定。然后,我们提出了一种基于深度基于学习的方法:充电控制深度确定性政策梯度(CDDPG),以了解满足用户对电池能量要求的最佳充电控制策略,同时最小化用户的充电费用。我们利用长短期内存(LSTM)网络提取以前的能量价格的信息来确定当前的充电控制策略。此外,高斯噪声被添加到演员网络的输出,以防止代理粘附到非优质策略中。此外,我们通过使用两个重放缓冲区来解决稀疏奖励的限制,其中一个用于在充电阶段存储奖励,另一个用于在充电完成后用于存储奖励。仿真结果证明了基于CDDPG的方法优于基于深度学习的方法(DQL)和基于深度确定的 - 政策梯度的方法(DDPG),以满足用户对电池能量的要求和减少充电成本。

著录项

  • 来源
    《Internet of Things Journal, IEEE》 |2021年第5期|3075-3087|共13页
  • 作者

    Zhang Feiye; Yang Qingyu; An Dou;

  • 作者单位

    Xi An Jiao Tong Univ Sch Automat Sci & Engn Xian 710049 Peoples R China;

    Xi An Jiao Tong Univ Sch Automat Sci & Engn Xian 710049 Peoples R China|Xi An Jiao Tong Univ MOE Key Lab Intelligent Networks & Network Secur SKLMSE Lab Xian 710049 Peoples R China;

    Xi An Jiao Tong Univ Sch Automat Sci & Engn Xian 710049 Peoples R China|Xi An Jiao Tong Univ MOE Key Lab Intelligent Networks & Network Secur SKLMSE Lab Xian 710049 Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Charging control; deep deterministic policy gradient (DDPG); electric vehicle (EV); Markov decision process (MDP);

    机译:充电控制;深度确定性政策梯度(DDPG);电动车(EV);马尔可夫决策过程(MDP);
  • 入库时间 2022-08-18 22:58:14

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号