Hamiltonian-Driven Adaptive Dynamic Programming With Efficient Experience Replay

IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS(2024)

引用 11|浏览44
暂无评分
摘要
This article presents a novel efficient experience-replay-based adaptive dynamic programming (ADP) for the optimal control problem of a class of nonlinear dynamical systems within the Hamiltonian-driven framework. The quasi-Hamiltonian is presented for the policy evaluation problem with an admissible policy. With the quasi-Hamiltonian, a novel composite critic learning mechanism is developed to combine the instantaneous data with the historical data. In addition, the pseudo-Hamiltonian is defined to deal with the performance optimization problem. Based on the pseudo-Hamiltonian, the conventional Hamilton-Jacobi-Bellman (HJB) equation can be represented in a filtered form, which can be implemented online. Theoretical analysis is investigated in terms of the convergence of the adaptive critic design and the stability of the closed-loop systems, where parameter convergence can be achieved under a weakened excitation condition. Simulation studies are investigated to verify the efficacy of the presented design scheme.
更多
查看译文
关键词
Mathematical models,Optimal control,Optimization,Convergence,Iterative algorithms,Dynamic programming,Learning systems,Hamilton-Jacobi-Bellman (HJB) equation,Hamiltonian-driven adaptive dynamic programming (ADP),pseudo-Hamiltonian,quasi-Hamiltonian,relaxed excitation condition
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要