Incorporating Online Learning Into MCTS-Based Intention Progression

Chengcheng Song,Yuan Yao,Sixian Chan

IEEE ACCESS(2024)

引用 0|浏览0
暂无评分
摘要
Agents have been applied to a wide variety of fields, including power systems and spacecraft. Belief-Desire-Intention (BDI) agents, as one of the most widely used and researched architectures, have the advantage of being able to pursue multiple goals in parallel. The problem of deciding "what to do" next at each of the agent's deliberation cycle is therefore critical for BDI agents, which is defined as the intention progression problem (IPP). Among all existing approaches to IPP, the majority of approaches have overlooked the significance of runtime historical data, thereby limiting the adaptability and decision-making capabilities of agents. In this paper, we propose to incorporate online learning into the current state-of-the-art intention progression approach S-A to overcome the above limitations. This approach not only prevents S-A from consuming computational resources on ineffective and inefficient simulations, but also significantly improves the execution efficiency of the agent. Especially when dealing with large-scale problem domains, this improvement significantly enhances the planning capability of the agents. In particular, we have proposed the SA(Q) and SA(L) schedulers, both of which can learn how to generate "reasonable" rollouts during the simulation phase of MCTS based on historical simulation data at run time. We compare the performance of our approach with the state-of-the-art S-A in a range of scenarios of increasing difficulty. The results demonstrate that our approaches outperform S-A , both in terms of the number of goals achieved and the computational overhead required.
更多
查看译文
关键词
BDI agents,intention progression problem,Monte-Carlo tree search,online learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要