Global Convergence Guarantees for Federated Policy Gradient Methods with Adversaries
arxiv(2024)
摘要
Federated Reinforcement Learning (FRL) allows multiple agents to
collaboratively build a decision making policy without sharing raw
trajectories. However, if a small fraction of these agents are adversarial, it
can lead to catastrophic results. We propose a policy gradient based approach
that is robust to adversarial agents which can send arbitrary values to the
server. Under this setting, our results form the first global convergence
guarantees with general parametrization. These results demonstrate resilience
with adversaries, while achieving sample complexity of order
𝒪̃( 1/ϵ^2( 1/N-f +
f^2/(N-f)^2)), where N is the total number of agents and
f is the number of adversarial agents.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要