Abstract
When a robot is learning it needs to explore its environment and how its environment responds on its actions. When the environment is large and there are a large number of possible actions the robot can take, this exploration phase can take prohibitively long. However, exploration can often be optimised by letting a human expert guide the robot during its learning. Interactive machine learning, in which a human user interactively guides the robot as it learns, has been shown to be an effective way to teach a robot. It requires an intuitive control mechanism to allow the human expert to provide feedback on the robot’s progress. This paper presents a novel method which combines Reinforcement Learning and Supervised Progressively Autonomous Robot Competencies (SPARC). By allowing the user to fully control the robot and by treating rewards as implicit, SPARC aims to learn an action policy while maintaining human supervisory oversight of the robot’s behaviour. This method is evaluated and compared to Interactive Reinforcement Learning in a robot teaching task. Qualitative and quantitative results indicate that SPARC allows for safer and faster learning by the robot, whilst not placing a high workload on the human teacher.
DOI
10.1016/j.patrec.2017.03.015
Publication Date
2017-03-18
Publication Title
Pattern Recognition Letters
ISSN
0167-8655
Embargo Period
2018-03-18
Organisational Unit
School of Engineering, Computing and Mathematics
Recommended Citation
Senft, E., Baxter, P., Kennedy, J., Lemaignan, S., & Belpaeme, T. (2017) 'Supervised autonomy for online learning in human-robot interaction', Pattern Recognition Letters, . Available at: https://doi.org/10.1016/j.patrec.2017.03.015