Abstract
Social interacting is a complex task for which machine learning holds particular promise. However, as no sufficiently accurate simulator of human interactions exists today, the learning of social interaction strategies has to happen online in the real world. Actions executed by the robot impact on humans, and as such have to be carefully selected, making it impossible to rely on random exploration. Additionally, no clear reward function exists for social interactions. This implies that traditional approaches used for Reinforcement Learning cannot be directly applied for learning how to interact with the social world. As such we argue that robots will profit from human expertise and guidance to learn social interactions. However, as the quantity of input a human can provide is limited, new methods have to be designed to use human input more efficiently. In this paper we describe a setup in which we combine a framework called Supervised Progressively Autonomous Robot Competencies (SPARC), which allows safer online learning with Reinforcement Learning, with the use of partial states radier than full states to accelerate generalisation and obtain a usable action policy more quickly.
Publication Date
2017-01-01
Publication Title
AAAI Fall Symposium - Technical Report
Volume
FS-17-01 - FS-17-05
Publisher
AAAI Press
ISBN
9781577357940
Embargo Period
2024-11-22
First Page
109
Last Page
113
Recommended Citation
Senft, E., Lemaignan, S., Baxter, P., & Belpaeme, T. (2017) 'Toward supervised reinforcement learning with partial states for social HRI', AAAI Fall Symposium - Technical Report, FS-17-01 - FS-17-05, pp. 109-113. AAAI Press: Retrieved from https://pearl.plymouth.ac.uk/secam-research/1637