Learning Causal Overhypotheses through Exploration in Children and Computational Models
Jiakun Liu
EECS Department, University of California, Berkeley
Technical Report No. UCB/EECS-2022-134
May 17, 2022
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-134.pdf
Despite recent progress in reinforcement learning (RL), RL algorithms for exploration still remain an active area of research. Existing methods often focus on state-based metrics, which do not consider the underlying causal structures of the environment, and while re- cent research has begun to explore RL environments for causal learning, these environments primarily leverage causal information through causal inference or induction rather than ex- ploration. In contrast, human children—some of the most proficient explorers—have been shown to use causal information to great benefit. In this work, we introduce a novel RL environment designed with a controllable causal structure, which allows us to evaluate ex- ploration strategies used by both agents and children in a unified environment. In addition, through experimentation on both computation models and children, we demonstrate that there are significant differences between information-gain optimal RL exploration in causal environments and the exploration of children in the same environments. We conclude with a discussion of how these findings may inspire new directions of research into efficient explo- ration and disambiguation of causal structures for RL algorithms.
Advisors: John F. Canny
BibTeX citation:
@mastersthesis{Liu:EECS-2022-134, Author= {Liu, Jiakun}, Title= {Learning Causal Overhypotheses through Exploration in Children and Computational Models}, School= {EECS Department, University of California, Berkeley}, Year= {2022}, Month= {May}, Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-134.html}, Number= {UCB/EECS-2022-134}, Abstract= {Despite recent progress in reinforcement learning (RL), RL algorithms for exploration still remain an active area of research. Existing methods often focus on state-based metrics, which do not consider the underlying causal structures of the environment, and while re- cent research has begun to explore RL environments for causal learning, these environments primarily leverage causal information through causal inference or induction rather than ex- ploration. In contrast, human children—some of the most proficient explorers—have been shown to use causal information to great benefit. In this work, we introduce a novel RL environment designed with a controllable causal structure, which allows us to evaluate ex- ploration strategies used by both agents and children in a unified environment. In addition, through experimentation on both computation models and children, we demonstrate that there are significant differences between information-gain optimal RL exploration in causal environments and the exploration of children in the same environments. We conclude with a discussion of how these findings may inspire new directions of research into efficient explo- ration and disambiguation of causal structures for RL algorithms.}, }
EndNote citation:
%0 Thesis %A Liu, Jiakun %T Learning Causal Overhypotheses through Exploration in Children and Computational Models %I EECS Department, University of California, Berkeley %D 2022 %8 May 17 %@ UCB/EECS-2022-134 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-134.html %F Liu:EECS-2022-134