posted on 2024-02-09, 17:56authored byEmmanuel Senft, Severin Lemaignan, Tony Belpaeme, Paul BaxterPaul Baxter
<p>Social interacting is a complex task for which machine learning holds particular promise. However, as no sufficiently accurate simulator of human interactions exists today, the learning of social interaction strategies has to happen online in the real world. Actions executed by the robot impact on humans, and as such have to be carefully selected, making it impossible to rely on random exploration. Additionally, no clear reward function exists for social interactions. This implies that traditional approaches used for Reinforcement Learning cannot be directly applied for learning how to interact with the social world. As such we argue that robots will profit from human expertise and guidance to learn social interactions. However, as the quantity of input a human can provide is limited, new methods have to be designed to use human input more efficiently. In this paper we describe a setup in which we combine a framework called Supervised Progressively Autonomous Robot Competencies (SPARC), which allows safer online learning with Reinforcement Learning, with the use of partial states rather than full states to accelerate generalisation and obtain a usable action policy more quickly.</p>
History
School affiliated with
School of Computer Science (Research Outputs)
Publisher
AAAI Press
Date Submitted
2018-02-02
Date Accepted
2017-11-09
Date of First Publication
2017-11-09
Date of Final Publication
2017-11-09
Event Name
4th AAAI FSS on Artificial Intelligence for Social Human-Robot Interaction (AI-HRI)