Compositional Models For Reinforcement Learning

MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT I(2009)

引用 7|浏览1
暂无评分
摘要
Innovations such as optimistic exploration, function approximation, and hierarchical decomposition have helped scale, reinforcement learning to more complex environments, but these three ideas have rarely been studied together. This paper develops a unified framework that formalizes these algorithmic contributions as, operators on learned models of the environment. Our formalism reveals some synergies among these innovations, and it, suggests a straight forward way to compose them. The resulting algorithm, Fitted R-MAXQ, is the first to combine the function approximation of fitted algorithms, the efficient; model-based exploration of R-MAX, and the hierarchical decompostion of MAXQ.
更多
查看译文
关键词
function approximation,efficient model-based exploration,hierarchical decomposition,hierarchical decompostion,optimistic exploration,Fitted R-MAXQ,algorithmic contribution,complex environment,fitted algorithm,resulting algorithm,Compositional Models,Reinforcement Learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要