Optimistic Active Learning of Task and Action Models for Robotic Manipulation

Manipulation tasks such as construction and assembly require reasoning over complex object interactions. In order to successfully plan for, execute, and achieve a given task, these interactions must be modeled accurately and capture low-level dynamics. Some examples include modeling how a constraine...

Full description

Bibliographic Details
Main Author: Moses, Caris
Other Authors: Lozano-Pérez, Tomás
Format: Thesis
Published: Massachusetts Institute of Technology 2022
Online Access:https://hdl.handle.net/1721.1/144676
https://orcid.org/0000-0002-6617-616X
_version_ 1811095451833729024
author Moses, Caris
author2 Lozano-Pérez, Tomás
author_facet Lozano-Pérez, Tomás
Moses, Caris
author_sort Moses, Caris
collection MIT
description Manipulation tasks such as construction and assembly require reasoning over complex object interactions. In order to successfully plan for, execute, and achieve a given task, these interactions must be modeled accurately and capture low-level dynamics. Some examples include modeling how a constrained object (such as a door) moves when grasped, the conditions under which an object will rest stably on another, or the friction constraints that allow an object to be pushed by another object. Acquiring models of object interactions for planning is a challenge. Existing engineering methods fail to accurately capture how an object’s properties such as friction, shape, and mass distribution, effect the success of actions such as pushing and stacking. Therefore, in this work we leverage machine learning as a data-driven approach to acquiring action models, with the hope that one day a robot equipped with a learning strategy and some basic understanding of the world could learn composable action models useful for planning to achieve a myriad of tasks. We see this work as a small step in this direction. Acquiring accurate models through a data-driven approach requires the robot to conduct a vast amount of information-rich interactions in the world. Collecting data on both real and simulated platforms can be time and cost prohibitive. In this work we take an active learning approach to aid the robot in finding the small subspace of informative actions within the large action space it has available to explore (all motions, grasps, and object interactions). Additionally, we supply the robot with optimistic action models, which are a relaxation of the true dynamics models. These models provide structure by constraining the exploration space in order to improve learning efficiency. Optimistic action models have the additional benefit of being easier to specify than fully accurate action models. We are generally interested in the scenario in which a robot is given an initial (optimistic) action model, an active learning strategy, and a space of domain-specific problems to generalize over. First, we give a method for learning task models in a bandit problem setting for constrained mechanisms. Our method, Contextual Prior Prediction, enables quick task success at evaluation time through the use of a learned vision-based prior. Then, we give a novel active learning strategy, Sequential Actions, for learning action models for long-horizon manipulation tasks in a block stacking domain. Finally, we give results in a tool use domain for our Sequential Goals method which improves upon Sequential Actions by exploring goal-directed plans at training time.
first_indexed 2024-09-23T16:16:59Z
format Thesis
id mit-1721.1/144676
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T16:16:59Z
publishDate 2022
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1446762022-08-30T03:04:21Z Optimistic Active Learning of Task and Action Models for Robotic Manipulation Moses, Caris Lozano-Pérez, Tomás Kaelbling, Leslie Pack Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Manipulation tasks such as construction and assembly require reasoning over complex object interactions. In order to successfully plan for, execute, and achieve a given task, these interactions must be modeled accurately and capture low-level dynamics. Some examples include modeling how a constrained object (such as a door) moves when grasped, the conditions under which an object will rest stably on another, or the friction constraints that allow an object to be pushed by another object. Acquiring models of object interactions for planning is a challenge. Existing engineering methods fail to accurately capture how an object’s properties such as friction, shape, and mass distribution, effect the success of actions such as pushing and stacking. Therefore, in this work we leverage machine learning as a data-driven approach to acquiring action models, with the hope that one day a robot equipped with a learning strategy and some basic understanding of the world could learn composable action models useful for planning to achieve a myriad of tasks. We see this work as a small step in this direction. Acquiring accurate models through a data-driven approach requires the robot to conduct a vast amount of information-rich interactions in the world. Collecting data on both real and simulated platforms can be time and cost prohibitive. In this work we take an active learning approach to aid the robot in finding the small subspace of informative actions within the large action space it has available to explore (all motions, grasps, and object interactions). Additionally, we supply the robot with optimistic action models, which are a relaxation of the true dynamics models. These models provide structure by constraining the exploration space in order to improve learning efficiency. Optimistic action models have the additional benefit of being easier to specify than fully accurate action models. We are generally interested in the scenario in which a robot is given an initial (optimistic) action model, an active learning strategy, and a space of domain-specific problems to generalize over. First, we give a method for learning task models in a bandit problem setting for constrained mechanisms. Our method, Contextual Prior Prediction, enables quick task success at evaluation time through the use of a learned vision-based prior. Then, we give a novel active learning strategy, Sequential Actions, for learning action models for long-horizon manipulation tasks in a block stacking domain. Finally, we give results in a tool use domain for our Sequential Goals method which improves upon Sequential Actions by exploring goal-directed plans at training time. Ph.D. 2022-08-29T16:03:59Z 2022-08-29T16:03:59Z 2022-05 2022-06-21T19:15:15.418Z Thesis https://hdl.handle.net/1721.1/144676 https://orcid.org/0000-0002-6617-616X In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Moses, Caris
Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title_full Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title_fullStr Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title_full_unstemmed Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title_short Optimistic Active Learning of Task and Action Models for Robotic Manipulation
title_sort optimistic active learning of task and action models for robotic manipulation
url https://hdl.handle.net/1721.1/144676
https://orcid.org/0000-0002-6617-616X
work_keys_str_mv AT mosescaris optimisticactivelearningoftaskandactionmodelsforroboticmanipulation