Automated Discovery of Functional Actual Causes in Complex Environments
arxiv(2024)
摘要
Reinforcement learning (RL) algorithms often struggle to learn policies that
generalize to novel situations due to issues such as causal confusion,
overfitting to irrelevant factors, and failure to isolate control of state
factors. These issues stem from a common source: a failure to accurately
identify and exploit state-specific causal relationships in the environment.
While some prior works in RL aim to identify these relationships explicitly,
they rely on informal domain-specific heuristics such as spatial and temporal
proximity. Actual causality offers a principled and general framework for
determining the causes of particular events. However, existing definitions of
actual cause often attribute causality to a large number of events, even if
many of them rarely influence the outcome. Prior work on actual causality
proposes normality as a solution to this problem, but its existing
implementations are challenging to scale to complex and continuous-valued RL
environments. This paper introduces functional actual cause (FAC), a framework
that uses context-specific independencies in the environment to restrict the
set of actual causes. We additionally introduce Joint Optimization for Actual
Cause Inference (JACI), an algorithm that learns from observational data to
infer functional actual causes. We demonstrate empirically that FAC agrees with
known results on a suite of examples from the actual causality literature, and
JACI identifies actual causes with significantly higher accuracy than existing
heuristic methods in a set of complex, continuous-valued environments.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要