Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC

The deployment of agile autonomous systems in challenging, unstructured environments requires adaptation capabilities and robustness to uncertainties. Existing robust and adaptive controllers, such as those based on model predictive control (MPC), can achieve impressive performance at the cost of he...

Full description

Bibliographic Details
Main Author: Zhao, Tong
Other Authors: How, Jonathan P.
Format: Thesis
Published: Massachusetts Institute of Technology 2023
Online Access:https://hdl.handle.net/1721.1/152818
_version_ 1811082214998278144
author Zhao, Tong
author2 How, Jonathan P.
author_facet How, Jonathan P.
Zhao, Tong
author_sort Zhao, Tong
collection MIT
description The deployment of agile autonomous systems in challenging, unstructured environments requires adaptation capabilities and robustness to uncertainties. Existing robust and adaptive controllers, such as those based on model predictive control (MPC), can achieve impressive performance at the cost of heavy online onboard computations. Strategies that efficiently learn robust and onboard-deployable policies from MPC have emerged, but they still lack fundamental adaptation capabilities. In this work, we extend an existing efficient Imitation Learning (IL) algorithm for robust policy learning from MPC with the ability to learn policies that adapt to challenging model/environment uncertainties. The key idea of our approach consists of modifying the IL procedure by conditioning the policy on a learned lower-dimensional model/environment representation that can be efficiently estimated online. We tailor our approach to learning an adaptive position and attitude control policy to track trajectories under challenging disturbances on a multirotor. Our evaluation shows that a high-quality adaptive policy can be obtained in about 1.3 hours of combined demonstration and training time. We empirically demonstrate rapid adaptation to in- and out-of-training-distribution uncertainties, achieving a 6.1 cm average position error under wind disturbances that correspond to 50% of the weight of the robot, and that are 36% larger than the maximum wind seen during training. Additionally, we verify the performance of our controller during real-world deployment in multiple trajectories, demonstrating adaptation to turbulent winds of up to 5.2 m/s and slung loads of up to 40% of the robot’s mass, and reducing the average position error on each trajectory to under 15 cm, a 70% improvement compared to a non-adaptive baseline.
first_indexed 2024-09-23T11:59:32Z
format Thesis
id mit-1721.1/152818
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T11:59:32Z
publishDate 2023
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1528182023-11-03T03:37:39Z Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC Zhao, Tong How, Jonathan P. Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science The deployment of agile autonomous systems in challenging, unstructured environments requires adaptation capabilities and robustness to uncertainties. Existing robust and adaptive controllers, such as those based on model predictive control (MPC), can achieve impressive performance at the cost of heavy online onboard computations. Strategies that efficiently learn robust and onboard-deployable policies from MPC have emerged, but they still lack fundamental adaptation capabilities. In this work, we extend an existing efficient Imitation Learning (IL) algorithm for robust policy learning from MPC with the ability to learn policies that adapt to challenging model/environment uncertainties. The key idea of our approach consists of modifying the IL procedure by conditioning the policy on a learned lower-dimensional model/environment representation that can be efficiently estimated online. We tailor our approach to learning an adaptive position and attitude control policy to track trajectories under challenging disturbances on a multirotor. Our evaluation shows that a high-quality adaptive policy can be obtained in about 1.3 hours of combined demonstration and training time. We empirically demonstrate rapid adaptation to in- and out-of-training-distribution uncertainties, achieving a 6.1 cm average position error under wind disturbances that correspond to 50% of the weight of the robot, and that are 36% larger than the maximum wind seen during training. Additionally, we verify the performance of our controller during real-world deployment in multiple trajectories, demonstrating adaptation to turbulent winds of up to 5.2 m/s and slung loads of up to 40% of the robot’s mass, and reducing the average position error on each trajectory to under 15 cm, a 70% improvement compared to a non-adaptive baseline. M.Eng. 2023-11-02T20:18:45Z 2023-11-02T20:18:45Z 2023-09 2023-10-03T18:21:17.820Z Thesis https://hdl.handle.net/1721.1/152818 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Zhao, Tong
Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title_full Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title_fullStr Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title_full_unstemmed Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title_short Efficiently Learning Robust, Adaptive Controllers from Robust Tube MPC
title_sort efficiently learning robust adaptive controllers from robust tube mpc
url https://hdl.handle.net/1721.1/152818
work_keys_str_mv AT zhaotong efficientlylearningrobustadaptivecontrollersfromrobusttubempc