Learning structured representations for perception and control

Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, 2016.

Bibliographic Details
Main Author: Kulkarni, Tejas Dattatraya
Other Authors: Joshua B. Tenenbaum.
Format: Thesis
Language:eng
Published: Massachusetts Institute of Technology 2017
Subjects:
Online Access:http://hdl.handle.net/1721.1/107557
_version_ 1826189884991209472
author Kulkarni, Tejas Dattatraya
author2 Joshua B. Tenenbaum.
author_facet Joshua B. Tenenbaum.
Kulkarni, Tejas Dattatraya
author_sort Kulkarni, Tejas Dattatraya
collection MIT
description Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, 2016.
first_indexed 2024-09-23T08:27:21Z
format Thesis
id mit-1721.1/107557
institution Massachusetts Institute of Technology
language eng
last_indexed 2024-09-23T08:27:21Z
publishDate 2017
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1075572019-04-09T18:17:42Z Learning structured representations for perception and control Kulkarni, Tejas Dattatraya Joshua B. Tenenbaum. Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences. Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences. Brain and Cognitive Sciences. Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, 2016. Cataloged from PDF version of thesis. Includes bibliographical references (pages 117-129). I argue that the intersection of deep learning, hierarchical reinforcement learning, and generative models provides a promising avenue towards building agents that learn to produce goal-directed behavior given sensations. I present models and algorithms that learn from raw observations and will emphasize on minimizing their sample complexity and number of training steps required for convergence. To this end, I introduce hierarchical variants of deep reinforcement learning algorithms, which produce and utilize temporally extended abstractions over actions. I also present a hybrid model-free and model-based deep reinforcement learning model, which can also be potentially used to automatically extract subgoals for bootstrapping temporal abstractions. I will then present a model-based approach for perception, which unifies deep learning and probabilistic models, to learn powerful representations of images without labeled data or external rewards. Learning goal-directed behavior with sparse and delayed rewards is a fundamental challenge for reinforcement learning algorithms. The primary difficulty arises due to insufficient exploration, resulting in an agent being unable to learn robust value functions. I present the Deep Hierarchical Reinforcement Learning (h-DQN) approach, which integrates hierarchical value functions operating at different time scales, along with goal-driven intrinsically motivated behavior for efficient exploration. Intrinsically motivated agents can explore new behavior for its own sake rather than to directly solve problems. Such intrinsic behaviors could eventually help the agent solve tasks posed by the environment. h-DQN allows for flexible goal specifications, such as functions over entities and relations. This provides an efficient space for exploration in complicated environments. I will demonstrate h-DQN's ability to learn optimal behavior given raw pixels in environments with very sparse and delayed feedback. I will then introduce the Deep Successor Reinforcement (DSR) learning approach. DSR is a hybrid model-free and model-based RL algorithm. It learns the value function of a state by taking the inner product between the state's expected future feature occupancy and the corresponding immediate rewards. This factorization of the value function has several appealing properties - increased sensitivity to changes in the reward structure and potentially the ability to automatically extract subgoals for learning temporal abstractions. Finally, I argue for the need for better representations of images, both in reinforcement learning tasks and in general. Existing deep learning approaches learn useful representations given lots of labeled data or rewards. Moreover, they also lack the inductive biases needed to disentangle causal structure in images such as objects, shape, pose and other intrinsic scene properties. I present generative models of vision, often referred to as analysis-by-synthesis approaches, by combining deep generative methods with probabilistic modeling. This approach aims to learn structured representations of images given raw observations. I argue that such intermediate representations will be crucial to scale-up deep reinforcement learning algorithms, and to bridge the gap between machine and human learning. by Tejas Dattatraya Kulkarni. Ph. D. 2017-03-20T19:39:55Z 2017-03-20T19:39:55Z 2016 2016 Thesis http://hdl.handle.net/1721.1/107557 974640245 eng MIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission. http://dspace.mit.edu/handle/1721.1/7582 129 pages application/pdf Massachusetts Institute of Technology
spellingShingle Brain and Cognitive Sciences.
Kulkarni, Tejas Dattatraya
Learning structured representations for perception and control
title Learning structured representations for perception and control
title_full Learning structured representations for perception and control
title_fullStr Learning structured representations for perception and control
title_full_unstemmed Learning structured representations for perception and control
title_short Learning structured representations for perception and control
title_sort learning structured representations for perception and control
topic Brain and Cognitive Sciences.
url http://hdl.handle.net/1721.1/107557
work_keys_str_mv AT kulkarnitejasdattatraya learningstructuredrepresentationsforperceptionandcontrol