Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents

Performance of state-of-the art offline and model-based reinforcement learning (RL) algorithms deteriorates significantly when subjected to severe data scarcity and the presence of heterogeneous agents. In this work, we propose a model-based offline RL method to approach this setting. Using all avai...

Full description

Bibliographic Details
Main Author: Alumootil, Varkey
Other Authors: Shah, Devavrat
Format: Thesis
Published: Massachusetts Institute of Technology 2022
Online Access:https://hdl.handle.net/1721.1/139143
_version_ 1826192706900066304
author Alumootil, Varkey
author2 Shah, Devavrat
author_facet Shah, Devavrat
Alumootil, Varkey
author_sort Alumootil, Varkey
collection MIT
description Performance of state-of-the art offline and model-based reinforcement learning (RL) algorithms deteriorates significantly when subjected to severe data scarcity and the presence of heterogeneous agents. In this work, we propose a model-based offline RL method to approach this setting. Using all available data from the various agents, we construct personalized simulators for each individual agent, which are then used to train RL policies. We do so by modeling the transition dynamics of the agents as a low rank tensor decomposition of latent factors associated with agents, states, and actions. We perform experiments on various benchmark environments and demonstrate improvement over existing offline approaches in the scarce data regime.
first_indexed 2024-09-23T09:28:01Z
format Thesis
id mit-1721.1/139143
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T09:28:01Z
publishDate 2022
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1391432022-01-15T04:00:23Z Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents Alumootil, Varkey Shah, Devavrat Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Performance of state-of-the art offline and model-based reinforcement learning (RL) algorithms deteriorates significantly when subjected to severe data scarcity and the presence of heterogeneous agents. In this work, we propose a model-based offline RL method to approach this setting. Using all available data from the various agents, we construct personalized simulators for each individual agent, which are then used to train RL policies. We do so by modeling the transition dynamics of the agents as a low rank tensor decomposition of latent factors associated with agents, states, and actions. We perform experiments on various benchmark environments and demonstrate improvement over existing offline approaches in the scarce data regime. M.Eng. 2022-01-14T14:52:30Z 2022-01-14T14:52:30Z 2021-06 2021-06-17T20:12:49.354Z Thesis https://hdl.handle.net/1721.1/139143 In Copyright - Educational Use Permitted Copyright MIT http://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology
spellingShingle Alumootil, Varkey
Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title_full Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title_fullStr Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title_full_unstemmed Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title_short Data-Efficient Offline Reinforcement Learning with Heterogeneous Agents
title_sort data efficient offline reinforcement learning with heterogeneous agents
url https://hdl.handle.net/1721.1/139143
work_keys_str_mv AT alumootilvarkey dataefficientofflinereinforcementlearningwithheterogeneousagents