Gregory Kahn and Adam Villaflor and Bosen Ding and Pieter Abbeel and Sergey Levine

EECS Department, University of California, Berkeley

Technical Report No. UCB/EECS-2018-47

May 10, 2018

http://www2.eecs.berkeley.edu/Pubs/TechRpts/2018/EECS-2018-47.pdf

Enabling robots to autonomously navigate complex environments is essential for real-world deployment. Prior methods approach this problem by having the robot maintain an internal map of the world, and then use a localization and planning method to navigate through the internal map. However, these approaches often include a variety of assumptions, are computationally intensive, and do not learn from failures. In contrast, learning-based methods improve as the robot acts in the environment, but are difficult to deploy in the real-world due to their high sample complexity. To address the need to learn complex policies with few samples, we propose a generalized computation graph that subsumes value-based model-free methods and model-based methods, with specific instantiations interpolating between model-free and model-based. We then instantiate this graph to form a navigation model that learns from raw images and is sample efficient. Our simulated car experiments explore the design decisions of our navigation model, and show our approach outperforms single-step and N-step double Q-learning. We also evaluate our approach on a real-world RC car and show it can learn to navigate through a complex indoor environment with a few hours of fully autonomous, self-supervised training.

Advisors: Pieter Abbeel and Sergey Levine


BibTeX citation:

@mastersthesis{Kahn:EECS-2018-47,
    Author= {Kahn, Gregory and Villaflor, Adam and Ding, Bosen and Abbeel, Pieter and Levine, Sergey},
    Title= {Self-supervised Deep Reinforcement Learning with Generalized Computation Graphs for Robot Navigation},
    School= {EECS Department, University of California, Berkeley},
    Year= {2018},
    Month= {May},
    Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2018/EECS-2018-47.html},
    Number= {UCB/EECS-2018-47},
    Abstract= {Enabling robots to autonomously navigate complex environments is essential for real-world deployment. Prior methods approach this problem by having the robot maintain an internal map of the world, and then use a localization and planning method to navigate through the internal map. However, these approaches often include a variety of assumptions, are computationally intensive, and do not learn from failures. In contrast, learning-based methods improve as the robot acts in the environment, but are difficult to deploy in the real-world due to their high sample complexity. To address the need to learn complex policies with few samples, we propose a generalized computation graph that subsumes value-based model-free methods and model-based methods, with specific instantiations interpolating between model-free and model-based. We then instantiate this graph to form a navigation model that learns from raw images and is sample efficient. Our simulated car experiments explore the design decisions of our navigation model, and show our approach outperforms single-step and N-step double Q-learning. We also evaluate our approach on a real-world RC car and show it can learn to navigate through a complex indoor environment with a few hours of fully autonomous, self-supervised training.},
}

EndNote citation:

%0 Thesis
%A Kahn, Gregory 
%A Villaflor, Adam 
%A Ding, Bosen 
%A Abbeel, Pieter 
%A Levine, Sergey 
%T Self-supervised Deep Reinforcement Learning with Generalized Computation Graphs for Robot Navigation
%I EECS Department, University of California, Berkeley
%D 2018
%8 May 10
%@ UCB/EECS-2018-47
%U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2018/EECS-2018-47.html
%F Kahn:EECS-2018-47