Generalization in Deep Reinforcement Learning for Robotic Navigation by Reward Shaping

IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS(2024)

引用 0|浏览19
暂无评分
摘要
This paper addresses the application of Deep Reinforcement Learning (DRL) methods in the context of local navigation, i.e., a robot moves towards a goal location in unknown and cluttered workspaces equipped only with limited-range exteroceptive sensors. Collision avoidance policies based on DRL present advantages, but they are quite susceptible to local minima, once their capacity to learn suitable actions is limited to the sensor range. We address this issue by means of reward shaping in actor-critic networks. A dense reward function, that incorporates map information gained in the training stage, is proposed to increase the agent's capacity to decide about the best action. Also, we offer a comparison between the Twin Delayed Deep-Deterministic Policy Gradient (TD3) and Soft Actor-Critic (SAC) algorithms for training our policy. A set of sim-to-sim and sim-to-real trials illustrate that our proposed reward shaping outperforms the compared methods in terms of generalization, by arriving at the target at higher rates in maps that are prone to local minima and collisions.
更多
查看译文
关键词
Deep Reinforcement Learning,Local Navigation,Mobile Robots,Unknown Cluttered Environments
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要