Understanding representation learning for deep reinforcement learning

<p>Representation learning is essential to practical success of reinforcement learning. Through a state representation, an agent can describe its environment to efficiently explore the state space, generalize to new states and perform credit assignment from delayed feedback. These representati...

Full description

Bibliographic Details
Main Author: Le Lan, C
Other Authors: Bellemare, M
Format: Thesis
Language:English
Published: 2023
Subjects:
_version_ 1826311792541827072
author Le Lan, C
author2 Bellemare, M
author_facet Bellemare, M
Le Lan, C
author_sort Le Lan, C
collection OXFORD
description <p>Representation learning is essential to practical success of reinforcement learning. Through a state representation, an agent can describe its environment to efficiently explore the state space, generalize to new states and perform credit assignment from delayed feedback. These representations may be state abstractions, hand-engineered or fixed features or implied by a neural network. In this thesis, we investigate several desirable theoretical properties of state representations and, using this categorization, design novel principled RL algorithms aiming at learning these state representations at scale through deep learning.</p> <p>First, we consider state abstractions induced by behavioral metrics and their generalization properties. We show that supporting the continuity of the value function is central to generalization in reinforcement learning. Together with this formalization, we provide an empirical evaluation comparing various metrics and demonstrating the importance of the choice of a neighborhood in RL algorithms.</p> <p>Then, we draw on statistical learning theory to characterize what it means for arbitrary state features to generalize in RL. We introduce a new notion called effective dimension of a representation that drives the generalization to unseen states and demonstrate its usefulness for value-based deep reinforcement learning in Atari games.</p> <p>The third contribution of this dissertation is a scalable algorithm to learn a state representation from a very large number of auxiliary tasks through deep learning. It is a stochastic gradient descent method to learn the principal components of a target matrix by means of a neural network from a handful of entries.</p> <p>Finally, the last part presents our findings on how the state representation in reinforcement learning influences the quality of an agent’s predictions but is also shaped by these predictions. We provide a formal mathematical model for studying this phenomenon and show how these theoretical results can be leveraged to improve the quality of the learning process.</p>
first_indexed 2024-03-07T08:16:28Z
format Thesis
id oxford-uuid:fd4b8e37-83d6-4cd6-b069-6745dd35fc51
institution University of Oxford
language English
last_indexed 2024-03-07T08:16:28Z
publishDate 2023
record_format dspace
spelling oxford-uuid:fd4b8e37-83d6-4cd6-b069-6745dd35fc512024-01-03T11:41:06ZUnderstanding representation learning for deep reinforcement learningThesishttp://purl.org/coar/resource_type/c_db06uuid:fd4b8e37-83d6-4cd6-b069-6745dd35fc51Reinforcement learningEnglishHyrax Deposit2023Le Lan, CBellemare, MTeh, YWhiteson, S<p>Representation learning is essential to practical success of reinforcement learning. Through a state representation, an agent can describe its environment to efficiently explore the state space, generalize to new states and perform credit assignment from delayed feedback. These representations may be state abstractions, hand-engineered or fixed features or implied by a neural network. In this thesis, we investigate several desirable theoretical properties of state representations and, using this categorization, design novel principled RL algorithms aiming at learning these state representations at scale through deep learning.</p> <p>First, we consider state abstractions induced by behavioral metrics and their generalization properties. We show that supporting the continuity of the value function is central to generalization in reinforcement learning. Together with this formalization, we provide an empirical evaluation comparing various metrics and demonstrating the importance of the choice of a neighborhood in RL algorithms.</p> <p>Then, we draw on statistical learning theory to characterize what it means for arbitrary state features to generalize in RL. We introduce a new notion called effective dimension of a representation that drives the generalization to unseen states and demonstrate its usefulness for value-based deep reinforcement learning in Atari games.</p> <p>The third contribution of this dissertation is a scalable algorithm to learn a state representation from a very large number of auxiliary tasks through deep learning. It is a stochastic gradient descent method to learn the principal components of a target matrix by means of a neural network from a handful of entries.</p> <p>Finally, the last part presents our findings on how the state representation in reinforcement learning influences the quality of an agent’s predictions but is also shaped by these predictions. We provide a formal mathematical model for studying this phenomenon and show how these theoretical results can be leveraged to improve the quality of the learning process.</p>
spellingShingle Reinforcement learning
Le Lan, C
Understanding representation learning for deep reinforcement learning
title Understanding representation learning for deep reinforcement learning
title_full Understanding representation learning for deep reinforcement learning
title_fullStr Understanding representation learning for deep reinforcement learning
title_full_unstemmed Understanding representation learning for deep reinforcement learning
title_short Understanding representation learning for deep reinforcement learning
title_sort understanding representation learning for deep reinforcement learning
topic Reinforcement learning
work_keys_str_mv AT lelanc understandingrepresentationlearningfordeepreinforcementlearning