Human-in-the-Loop Reinforcement Learning for Adaptive Assistive Interfaces
Jensen Gao
EECS Department, University of California, Berkeley
Technical Report No. UCB/EECS-2022-62
May 11, 2022
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-62.pdf
Machine learning has shown great potential to facilitate more effective methods for human-computer interaction. This includes artificial intelligence-based interfaces that can assist users with performing their desired objectives with improved performance. In this technical report, we propose two human-in-the-loop deep reinforcement learning (RL) based methods to infer the intent of a user through only high-dimensional, noisy user inputs, while adapting to the user’s inputs and feedback over time, in order to assist the user in performing their desired objectives more effectively. In Chapter 1, we propose a deep RL approach that learns from human feedback for assistive typing interfaces, which we formulate as contextual bandit problems. In Chapter 2, we propose a method that extends this style of approach for robotics tasks, which require sequential decision making. We do this through leveraging autonomous pre-training with deep RL. We demonstrate the effectiveness of these approaches using simulated user inputs, real user studies where participants communicate intent through webcam eye gaze, and a pilot study using brain-computer interfaces.
Advisors: Sergey Levine
BibTeX citation:
@mastersthesis{Gao:EECS-2022-62, Author= {Gao, Jensen}, Title= {Human-in-the-Loop Reinforcement Learning for Adaptive Assistive Interfaces}, School= {EECS Department, University of California, Berkeley}, Year= {2022}, Month= {May}, Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-62.html}, Number= {UCB/EECS-2022-62}, Abstract= {Machine learning has shown great potential to facilitate more effective methods for human-computer interaction. This includes artificial intelligence-based interfaces that can assist users with performing their desired objectives with improved performance. In this technical report, we propose two human-in-the-loop deep reinforcement learning (RL) based methods to infer the intent of a user through only high-dimensional, noisy user inputs, while adapting to the user’s inputs and feedback over time, in order to assist the user in performing their desired objectives more effectively. In Chapter 1, we propose a deep RL approach that learns from human feedback for assistive typing interfaces, which we formulate as contextual bandit problems. In Chapter 2, we propose a method that extends this style of approach for robotics tasks, which require sequential decision making. We do this through leveraging autonomous pre-training with deep RL. We demonstrate the effectiveness of these approaches using simulated user inputs, real user studies where participants communicate intent through webcam eye gaze, and a pilot study using brain-computer interfaces.}, }
EndNote citation:
%0 Thesis %A Gao, Jensen %T Human-in-the-Loop Reinforcement Learning for Adaptive Assistive Interfaces %I EECS Department, University of California, Berkeley %D 2022 %8 May 11 %@ UCB/EECS-2022-62 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-62.html %F Gao:EECS-2022-62