A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals

Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019

Bibliographic Details
Main Author: Elkind, Daniel(Daniel Harris)
Other Authors: Adrien Verdelhan
Format: Thesis
Language:eng
Published: Massachusetts Institute of Technology 2020
Subjects:
Online Access:https://hdl.handle.net/1721.1/124585
_version_ 1826212671519719424
author Elkind, Daniel(Daniel Harris)
author2 Adrien Verdelhan
author_facet Adrien Verdelhan
Elkind, Daniel(Daniel Harris)
author_sort Elkind, Daniel(Daniel Harris)
collection MIT
description Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019
first_indexed 2024-09-23T15:32:28Z
format Thesis
id mit-1721.1/124585
institution Massachusetts Institute of Technology
language eng
last_indexed 2024-09-23T15:32:28Z
publishDate 2020
publisher Massachusetts Institute of Technology
record_format dspace
spelling mit-1721.1/1245852020-04-14T03:04:56Z A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals Elkind, Daniel(Daniel Harris) Adrien Verdelhan Sloan School of Management. Sloan School of Management Sloan School of Management. Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019 Cataloged from PDF version of thesis. Includes bibliographical references (pages 27-29). This paper focuses the optimal trading execution problem, where a trader seeks to maximize the proceeds from trading a given quantity of shares of a financial asset over a fixed-duration trading period, considering that trading impacts the future trajectory of prices. I propose a reinforcement learning (RL) algorithm to solve this maximization problem. I prove that the algorithm converges to the optimal solution in a large class of settings and point out a useful duality between the learning contraction and the dynamic programming PDE. Using simulations calibrated to historical exchange trading data, I show that (i) the algorithm reproduces the analytical solution for the case of random walk prices with a linear absolute price impact function and (ii) matches the output of classical dynamic programming methods for the case of geometric brownian motion prices with linear relative price impact. In the most relevant case, when a signal containing information about prices is introduced to the environment, traditional computational methods become intractable. My algorithm still finds the optimal execution policy, leading to a statistically and economically meaningful reduction in trading costs. by Daniel Elkind. S.M. in Management Research S.M.inManagementResearch Massachusetts Institute of Technology, Sloan School of Management 2020-04-13T18:28:51Z 2020-04-13T18:28:51Z 2019 2019 Thesis https://hdl.handle.net/1721.1/124585 1149013871 eng MIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission. http://dspace.mit.edu/handle/1721.1/7582 29 pages application/pdf Massachusetts Institute of Technology
spellingShingle Sloan School of Management.
Elkind, Daniel(Daniel Harris)
A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title_full A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title_fullStr A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title_full_unstemmed A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title_short A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
title_sort reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
topic Sloan School of Management.
url https://hdl.handle.net/1721.1/124585
work_keys_str_mv AT elkinddanieldanielharris areinforcementlearningalgorithmforefficientdynamictradingexecutioninthepresenceofsignals
AT elkinddanieldanielharris reinforcementlearningalgorithmforefficientdynamictradingexecutioninthepresenceofsignals