Self-Adaptive Driving in Nonstationary Environments through Conjectural Online Lookahead Adaptation

arxiv(2023)

引用 2|浏览9
暂无评分
摘要
Powered by deep representation learning, reinforcement learning (RL) provides an end-to-end learning framework capable of solving self-driving (SD) tasks without manual designs. However, time-varying nonstationary environments cause proficient but specialized RL policies to fail at execution time. For example, an RL-based SD policy trained under sunny days does not generalize well to rainy weather. Even though meta learning enables the RL agent to adapt to new tasks/environments, its offline operation fails to equip the agent with online adaptation ability when facing nonstationary environments. This work proposes an online meta reinforcement learning algorithm based on the \emph{conjectural online lookahead adaptation} (COLA). COLA determines the online adaptation at every step by maximizing the agent's conjecture of the future performance in a lookahead horizon. Experimental results demonstrate that under dynamically changing weather and lighting conditions, the COLA-based self-adaptive driving outperforms the baseline policies in terms of online adaptability. A demo video, source code, and appendixes are available at {\tt https://github.com/Panshark/COLA}
更多
查看译文
关键词
COLA-based self-adaptive driving,conjectural online lookahead adaptation,deep representation learning,end-to-end learning framework,execution time,meta learning,online adaptability,online adaptation ability,online meta reinforcement,re-inforcement learning,RL agent,RL-based SD policy,self-driving tasks,specialized RL policies,time-varying nonstationary environments
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要