Abstract

Social interacting is a complex task for which machine learning holds particular promise. However, as no sufficiently accurate simulator of human interactions exists today, the learning of social interaction strategies has to happen online in the real world. Actions executed by the robot impact on humans, and as such have to be carefully selected, making it impossible to rely on random exploration. Additionally, no clear reward function exists for social interactions. This implies that traditional approaches used for Reinforcement Learning cannot be directly applied for learning how to interact with the social world. As such we argue that robots will profit from human expertise and guidance to learn social interactions. However, as the quantity of input a human can provide is limited, new methods have to be designed to use human input more efficiently. In this paper we describe a setup in which we combine a framework called Supervised Progressively Autonomous Robot Competencies (SPARC), which allows safer online learning with Reinforcement Learning, with the use of partial states radier than full states to accelerate generalisation and obtain a usable action policy more quickly.

Publication Date

2017-01-01

Publication Title

AAAI Fall Symposium - Technical Report

Volume

FS-17-01 - FS-17-05

Publisher

AAAI Press

ISBN

9781577357940

Embargo Period

2024-11-22

First Page

109

Last Page

113

Share

COinS