Graph Diffusion Policy Optimization
CoRR(2024)
摘要
Recent research has made significant progress in optimizing diffusion models
for specific downstream objectives, which is an important pursuit in fields
such as graph generation for drug design. However, directly applying these
models to graph diffusion presents challenges, resulting in suboptimal
performance. This paper introduces graph diffusion policy optimization (GDPO),
a novel approach to optimize graph diffusion models for arbitrary (e.g.,
non-differentiable) objectives using reinforcement learning. GDPO is based on
an eager policy gradient tailored for graph diffusion models, developed through
meticulous analysis and promising improved performance. Experimental results
show that GDPO achieves state-of-the-art performance in various graph
generation tasks with complex and diverse objectives. Code is available at
https://github.com/sail-sg/GDPO.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要