lem of understanding causality just as well as they
can learn to decode the raw vision data fed to them
in Atari Games as in Mnih et al. and turn it into win-
ning strategies. Pearl (2000) argues that certain prob-
lems cannot be solved using correlation based statis-
tics alone; to progress to the second rung of his cau-
sation ladder, interventions need to be made. RL is
a learning framework which naturally performs inter-
ventions on the environment it seeks to learn about
yet its tools do automatically account for the mathe-
matical implications of making interventions.
The dog barometer problem that I present here is
deliberately simple and the state-space given to the
learner is not ideal for a learning algorithm. RL Meth-
ods do exist for settings with hidden variables and I
have not used them here. Expanding the state space
to include previous state values and actions may solve
the problem. However I do think that the problems
in RL raised by dog barometer are not simply the re-
sult of a straw-man argument. The presence of hidden
variables in real life learning applications is almost
certain as is the existence of non-trivial causal struc-
tures whose effect may linger over arbitrarily long
timescales thereby negating the efficacy of adding
more history. Every model of a real problem will be
misspecified to some extent; it is important to under-
stand when and why this matters. The fact that the
cognitive error is sufficiently common in social sci-
ence to be named Goodhart’s Law is a good indica-
tor that this is a policy failure case which is likely
to appear again and again in real life applications of
RL. In defence of RL, the performance of the A2C
algorithm even in the face of such misspecification is
very promising and warrants further investigation to
see whether this is consistent or an artefact of the en-
vironment.
Finally, I would like to continue to build an open
library of causal problems which new RL algorithms
can be benchmarked against. Such an approach using
the OpenAI interface has already benefited RL and I
think such a library will help widen the audience of
Causal RL to general RL researchers. In parallel it
would be useful to begin to build a taxonomy of cog-
nitive errors that AI suffers from, starting by investi-
gating whether others, similar to Campbell-Goodhart
can be recreated with RL.
ACKNOWLEDGEMENTS
This work is supported by an EPSRC PhD stu-
dentship.
REFERENCES
Bareinboim, E. (2020). Towards Causal Reinforcement
Learning (CRL). In Thirty-seventh International Con-
ference on Machine Learning (ICML2020).
Bareinboim, E. and Pearl, J. (2016). Causal inference and
the data-fusion problem. Proceedings of the National
Academy of Sciences of the United States of America,
113(27):7345–7352.
Buesing, L., Weber, T., Zwols, Y., Racaniere, S., Guez, A.,
Lespiau, J.-B., and Heess, N. (2019). Woulda, Coulda,
Shoulda: Counterfactually-guided policy search. In
ICLR.
Campbell, D. T. (1979). Assessing the impact of planned
social change. Evaluation and Program Planning,
2(1):67–90.
Campolo, A. and Crawford, K. (2020). Enchanted Deter-
minism: Power without Responsibility in Artificial In-
telligence. Engaging Science, Technology, and Soci-
ety, 6:1.
Gershman, S. J. (2015). Reinforcement learning and causal
models. Oxford Handbook of Causal Reasoning,
pages 1–32.
Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep
Learning.
Goodhart, C. A. E. (1984). Problems of Monetary Man-
agement: The UK Experience. Monetary Theory and
Practice, pages 91–121.
Guo, R., Cheng, L., Li, J., Hahn, P. R., and Liu, H. (2018).
A Survey of Learning Causality with Data: Prob-
lems and Methods. ACM Computing Surveys (CSUR),
53(4):1–37.
Ha, D. and Schmidhuber, J. (2018). World Models.
Lehman, J., Clune, J., and Misevic, D. (2020). The sur-
prising creativity of digital evolution: A collection
of anecdotes from the evolutionary computation and
artificial life research communities. Artificial Life,
26(2):274–306.
Manheim, D. and Garrabrant, S. (2018). Categorizing Vari-
ants of Goodhart’s Law.
Mnih, V., Badia, A. P., Mirza, L., Graves, A., Harley,
T., Lillicrap, T. P., Silver, D., and Kavukcuoglu, K.
(2016). Asynchronous methods for deep reinforce-
ment learning. 33rd International Conference on Ma-
chine Learning, ICML 2016, 4:2850–2869.
Mnih, V., Kavukcuoglu, K., Silver, D., ..., and Hassabis, D.
(2015). Human-level control through deep reinforce-
ment learning. Nature, 518(7540):529–533.
Pearl, J. (2000). Causality: Models, reasoning and infer-
ence. Cambridge University Press.
Pearl, J. and Mackensie, D. (2018). The Book of Why: The
new science of cause and effect. Basic Books.
Rezende, D. J., Danihelka, I., Papamakarios, G., ..., and
Buesing, L. (2020). Causally Correct Partial Models
for Reinforcement Learning.
Rodamar, J. (2018). There ought to be a law! Campbell
versus Goodhart. Significance, 15(6):9.
Silver, D., Schrittwieser, J., Simonyan, K., ..., and Hassabis,
D. (2017). Mastering the game of Go without human
knowledge. Nature, 550(7676):354–359.
ICAART 2021 - 13th International Conference on Agents and Artificial Intelligence
72