Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks

We present a framework for automatically learning human user models from joint-action demonstrations that enables a robot to compute a robust policy for a collaborative task with a human. First, the demonstrated action sequences are clustered into different human types using an unsupervised learning...

Full description

Bibliographic Details
Main Authors: Shah, Julie A, Nikolaidis, Stefanos, Ramakrishnan, Ramya, Gu, Keren
Other Authors: Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Format: Article
Language:en_US
Published: Institute of Electrical and Electronics Engineers (IEEE) 2017
Online Access:http://hdl.handle.net/1721.1/107887
https://orcid.org/0000-0003-1338-8107
https://orcid.org/0000-0001-8239-5963
_version_ 1811086140413837312
author Shah, Julie A
Nikolaidis, Stefanos
Ramakrishnan, Ramya
Gu, Keren
author2 Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
author_facet Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory
Shah, Julie A
Nikolaidis, Stefanos
Ramakrishnan, Ramya
Gu, Keren
author_sort Shah, Julie A
collection MIT
description We present a framework for automatically learning human user models from joint-action demonstrations that enables a robot to compute a robust policy for a collaborative task with a human. First, the demonstrated action sequences are clustered into different human types using an unsupervised learning algorithm. A reward function is then learned for each type through the employment of an inverse reinforcement learning algorithm. The learned model is then incorporated into a mixed-observability Markov decision process (MOMDP) formulation, wherein the human type is a partially observable variable. With this framework, we can infer online the human type of a new user that was not included in the training set, and can compute a policy for the robot that will be aligned to the preference of this user. In a human subject experiment (n=30), participants agreed more strongly that the robot anticipated their actions when working with a robot incorporating the proposed framework (p<0.01), compared to manually annotating robot actions. In trials where participants faced difficulty annotating the robot actions to complete the task, the proposed framework significantly improved team efficiency (p<0.01). The robot incorporating the framework was also found to be more responsive to human actions compared to policies computed using a hand-coded reward function by a domain expert (p<0.01). These results indicate that learning human user models from joint-action demonstrations and encoding them in a MOMDP formalism can support effective teaming in human-robot collaborative tasks.
first_indexed 2024-09-23T13:21:29Z
format Article
id mit-1721.1/107887
institution Massachusetts Institute of Technology
language en_US
last_indexed 2024-09-23T13:21:29Z
publishDate 2017
publisher Institute of Electrical and Electronics Engineers (IEEE)
record_format dspace
spelling mit-1721.1/1078872022-09-28T13:39:21Z Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks Shah, Julie A Nikolaidis, Stefanos Ramakrishnan, Ramya Gu, Keren Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology. Department of Aeronautics and Astronautics Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Shah, Julie A Nikolaidis, Stefanos Ramakrishnan, Ramya Gu, Keren We present a framework for automatically learning human user models from joint-action demonstrations that enables a robot to compute a robust policy for a collaborative task with a human. First, the demonstrated action sequences are clustered into different human types using an unsupervised learning algorithm. A reward function is then learned for each type through the employment of an inverse reinforcement learning algorithm. The learned model is then incorporated into a mixed-observability Markov decision process (MOMDP) formulation, wherein the human type is a partially observable variable. With this framework, we can infer online the human type of a new user that was not included in the training set, and can compute a policy for the robot that will be aligned to the preference of this user. In a human subject experiment (n=30), participants agreed more strongly that the robot anticipated their actions when working with a robot incorporating the proposed framework (p<0.01), compared to manually annotating robot actions. In trials where participants faced difficulty annotating the robot actions to complete the task, the proposed framework significantly improved team efficiency (p<0.01). The robot incorporating the framework was also found to be more responsive to human actions compared to policies computed using a hand-coded reward function by a domain expert (p<0.01). These results indicate that learning human user models from joint-action demonstrations and encoding them in a MOMDP formalism can support effective teaming in human-robot collaborative tasks. 2017-04-05T20:03:20Z 2017-04-05T20:03:20Z 2015-03 Article http://purl.org/eprint/type/ConferencePaper 9781450328838 http://hdl.handle.net/1721.1/107887 Nikolaidis, Stefanos et al. “Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks.” ACM Press, 2015. 189–196. https://orcid.org/0000-0003-1338-8107 https://orcid.org/0000-0001-8239-5963 en_US http://dx.doi.org/10.1145/2696454.2696455 Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction - HRI '15 Creative Commons Attribution-Noncommercial-Share Alike http://creativecommons.org/licenses/by-nc-sa/4.0/ application/pdf Institute of Electrical and Electronics Engineers (IEEE) MIT web domain
spellingShingle Shah, Julie A
Nikolaidis, Stefanos
Ramakrishnan, Ramya
Gu, Keren
Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title_full Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title_fullStr Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title_full_unstemmed Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title_short Efficient Model Learning from Joint-Action Demonstrations for Human-Robot Collaborative Tasks
title_sort efficient model learning from joint action demonstrations for human robot collaborative tasks
url http://hdl.handle.net/1721.1/107887
https://orcid.org/0000-0003-1338-8107
https://orcid.org/0000-0001-8239-5963
work_keys_str_mv AT shahjuliea efficientmodellearningfromjointactiondemonstrationsforhumanrobotcollaborativetasks
AT nikolaidisstefanos efficientmodellearningfromjointactiondemonstrationsforhumanrobotcollaborativetasks
AT ramakrishnanramya efficientmodellearningfromjointactiondemonstrationsforhumanrobotcollaborativetasks
AT gukeren efficientmodellearningfromjointactiondemonstrationsforhumanrobotcollaborativetasks