谷歌Chrome浏览器插件
订阅小程序
在清言上使用

Approximate Dynamic Programming with Enhanced Off-policy Learning for Coordinating Distributed Energy Resources

Avijit Das,Di Wu,Bilal Ahmad Bhatti, Mohamed Kamaludeen

IEEE Transactions on Sustainable Energy(2024)

引用 0|浏览4
暂无评分
摘要
This paper proposes an innovative approximate dynamic programming (ADP) method for distributed energy resource coordination with the loss of life of battery energy storage system (BESS) explicitly modeled. The dispatch policy is designed to account for both calendrical and cyclical aging effects on BESS, explicitly modeling the impacts of ambient temperature on BESS lifespan. The proposed ADP employs an adaptive critic method and enhanced off-policy deterministic policy gradient (DPG) strategy, addressing the limitations of the on-policy gradient-based ADP approaches, including inadequate exploration, low data usage, and computational complexity. In particular, a customized policy is proposed to guide the algorithm to explore some promising decisions and thereby improve exploration capability and learning efficiency compared to conventional DPG-based learning approaches, which may struggle to find a global optimum due to random noisy action-based exploration or require expert demonstration with extra effort. The proposed method is illustrated using the IEEE 123-node system and compared with the existing ADP methods to prove solution accuracy and demonstrate the effects of incorporating degradation models into control design. Case studies showed that the proposed ADP effectively coordinates DERs with a 10 times smaller optimization gap compared to existing methods, and the incorporation of the BESS life loss model into the proposed control ensures the expected lifespan and results in significant cost savings.
更多
查看译文
关键词
Approximate dynamic programming,customized exploration,deterministic policy gradient,distributed energy resources,energy storage
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要