首页> 外文会议>Power Systems Computation Conference >Energy management of PV-storage systems: ADP approach with temporal difference learning
【24h】

Energy management of PV-storage systems: ADP approach with temporal difference learning

机译:光伏存储系统的能源管理:具有时差学习功能的ADP方法

获取原文

摘要

In the future, residential energy users can seize the full potential of demand response schemes by using an automated home energy management system (HEMS) to schedule their distributed energy resources. In order to generate high quality schedules, a HEMS needs to consider the stochastic nature of the PV generation and energy consumption as well as its inter-daily variations over several days. However, extending the decision horizon of proposed optimisation techniques is computationally difficult and moreover, these approaches are only computationally feasible with a limited number of storage devices and a low-resolution decision horizon. Given these existing shortcomings, this paper presents an approximate dynamic programming (ADP) approach with temporal difference learning for implementing a computationally efficient HEMS. In ADP, we obtain policies from value function approximations by stepping forward in time, compared to the value functions obtained by backward induction in DP. We use empirical data collected during the Smart Grid Smart City project in NSW, Australia, to estimate the parameters of a Markov chain model of PV output and electrical demand, which are then used in all simulations. To evaluate the quality of the solutions generated by ADP, we compare the ADP method to stochastic mixed-integer linear programming (MILP) and dynamic programming (DP). Our results show that ADP computes a solution much quicker than both DP and stochastic MILP, while providing better quality solutions than stochastic MILP and only a slight reduction in quality compared to the DP solution. Moreover, unlike the computationally-intensive DP, the ADP approach is able to consider a decision horizon beyond one day while also considering multiple storage devices, which results in a HEMS that can capture additional financial benefits
机译:将来,住宅能源用户可以通过使用自动家庭能源管理系统(HEMS)来调度其分布式能源,从而充分利用需求响应计划的潜力。为了生成高质量的计划,HEMS需要考虑光伏发电和能源消耗的随机性,以及其几天内的日间变化。然而,扩展所提出的优化技术的决策视野在计算上是困难的,此外,这些方法仅在有限数量的存储设备和低分辨率决策视野的情况下在计算上是可行的。鉴于这些现有缺陷,本文提出了一种具有时差学习的近似动态编程(ADP)方法,以实现计算效率高的HEMS。在ADP中,与在DP中通过反向归纳所获得的价值函数相比,我们通过在时间上向前迈进来从价值函数近似中获得策略。我们使用在澳大利亚新南威尔士州的“智能电网智能城市”项目中收集的经验数据来估算光伏输出和电力需求的马尔可夫链模型的参数,然后将其用于所有模拟中。为了评估由ADP生成的解决方案的质量,我们将ADP方法与随机混合整数线性规划(MILP)和动态规划(DP)进行了比较。我们的结果表明,ADP计算解决方案比DP和随机MILP都快得多,同时提供的质量比随机MILP更好,而与DP解决方案相比质量仅稍有下降。此外,与计算量大的DP不同,ADP方法能够考虑一天以上的决策范围,同时还考虑多个存储设备,这使得HEMS可以捕获额外的财务收益

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号