Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation

What is the “state” of a pile of objects? Collections of standard Lagrangian states ideally describe a pile, but it is impractical to estimate Lagrangian states directly from images to use for visuomotor feedback control. Given this burden of state estimation, is there a practical alternative repres...

Full description

Bibliographic Details
Main Author: Suh, Hyung Ju Terry
Other Authors: Tedrake, Russell L.
Format: Thesis
Published: Massachusetts Institute of Technology 2022
Online Access:https://hdl.handle.net/1721.1/143378
_version_ 1826212834401320960
author Suh, Hyung Ju Terry
author2 Tedrake, Russell L.
author_facet Tedrake, Russell L.
Suh, Hyung Ju Terry
author_sort Suh, Hyung Ju Terry
collection MIT
description What is the “state” of a pile of objects? Collections of standard Lagrangian states ideally describe a pile, but it is impractical to estimate Lagrangian states directly from images to use for visuomotor feedback control. Given this burden of state estimation, is there a practical alternative representation that lies closer to our observations? In addition, how can we build predictive models over such representations that can be useful for their task-free generality? In the first chapter of this thesis, we investigate using the image observation directly as state, and compare different models that can be useful over this space of representations. We surprisingly find that completely linear models that describe the evolution of images outperform naive deep models, and perform in par with models that work over particle-space representations. In the next chapter, we analyze and describe the reason for this inductive bias of linear models by describing the pixel space as a space of measures, and show limitations of this approach outside of object pile manipulation. In the final chapter of this thesis, we present a more general solution to image-based control based on doing model-based Reinforcement Learning on the sufficient statistics of a task, which we call Approximate Information States (AIS). We demonstrate that when the model does not have sufficient inductive bias, model-based reinforcement learning is prone to two important pitfalls: distribution shift, and optimization exploiting model error. These problems are tackled through online learning, and risk-aware control that penalizes the variance of the model ensemble.
first_indexed 2024-09-23T15:38:58Z
format Thesis
id mit-1721.1/143378
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T15:38:58Z
publishDate 2022
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1433782022-06-16T03:08:08Z Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation Suh, Hyung Ju Terry Tedrake, Russell L. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science What is the “state” of a pile of objects? Collections of standard Lagrangian states ideally describe a pile, but it is impractical to estimate Lagrangian states directly from images to use for visuomotor feedback control. Given this burden of state estimation, is there a practical alternative representation that lies closer to our observations? In addition, how can we build predictive models over such representations that can be useful for their task-free generality? In the first chapter of this thesis, we investigate using the image observation directly as state, and compare different models that can be useful over this space of representations. We surprisingly find that completely linear models that describe the evolution of images outperform naive deep models, and perform in par with models that work over particle-space representations. In the next chapter, we analyze and describe the reason for this inductive bias of linear models by describing the pixel space as a space of measures, and show limitations of this approach outside of object pile manipulation. In the final chapter of this thesis, we present a more general solution to image-based control based on doing model-based Reinforcement Learning on the sufficient statistics of a task, which we call Approximate Information States (AIS). We demonstrate that when the model does not have sufficient inductive bias, model-based reinforcement learning is prone to two important pitfalls: distribution shift, and optimization exploiting model error. These problems are tackled through online learning, and risk-aware control that penalizes the variance of the model ensemble. S.M. 2022-06-15T13:16:30Z 2022-06-15T13:16:30Z 2022-02 2022-03-04T20:59:52.218Z Thesis https://hdl.handle.net/1721.1/143378 In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Suh, Hyung Ju Terry
Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title_full Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title_fullStr Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title_full_unstemmed Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title_short Predictive Models for Visuomotor Feedback Control in Object Pile Manipulation
title_sort predictive models for visuomotor feedback control in object pile manipulation
url https://hdl.handle.net/1721.1/143378
work_keys_str_mv AT suhhyungjuterry predictivemodelsforvisuomotorfeedbackcontrolinobjectpilemanipulation