A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals
Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Language: | eng |
Published: |
Massachusetts Institute of Technology
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/1721.1/124585 |
_version_ | 1826212671519719424 |
---|---|
author | Elkind, Daniel(Daniel Harris) |
author2 | Adrien Verdelhan |
author_facet | Adrien Verdelhan Elkind, Daniel(Daniel Harris) |
author_sort | Elkind, Daniel(Daniel Harris) |
collection | MIT |
description | Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019 |
first_indexed | 2024-09-23T15:32:28Z |
format | Thesis |
id | mit-1721.1/124585 |
institution | Massachusetts Institute of Technology |
language | eng |
last_indexed | 2024-09-23T15:32:28Z |
publishDate | 2020 |
publisher | Massachusetts Institute of Technology |
record_format | dspace |
spelling | mit-1721.1/1245852020-04-14T03:04:56Z A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals Elkind, Daniel(Daniel Harris) Adrien Verdelhan Sloan School of Management. Sloan School of Management Sloan School of Management. Thesis: S.M. in Management Research, Massachusetts Institute of Technology, Sloan School of Management, 2019 Cataloged from PDF version of thesis. Includes bibliographical references (pages 27-29). This paper focuses the optimal trading execution problem, where a trader seeks to maximize the proceeds from trading a given quantity of shares of a financial asset over a fixed-duration trading period, considering that trading impacts the future trajectory of prices. I propose a reinforcement learning (RL) algorithm to solve this maximization problem. I prove that the algorithm converges to the optimal solution in a large class of settings and point out a useful duality between the learning contraction and the dynamic programming PDE. Using simulations calibrated to historical exchange trading data, I show that (i) the algorithm reproduces the analytical solution for the case of random walk prices with a linear absolute price impact function and (ii) matches the output of classical dynamic programming methods for the case of geometric brownian motion prices with linear relative price impact. In the most relevant case, when a signal containing information about prices is introduced to the environment, traditional computational methods become intractable. My algorithm still finds the optimal execution policy, leading to a statistically and economically meaningful reduction in trading costs. by Daniel Elkind. S.M. in Management Research S.M.inManagementResearch Massachusetts Institute of Technology, Sloan School of Management 2020-04-13T18:28:51Z 2020-04-13T18:28:51Z 2019 2019 Thesis https://hdl.handle.net/1721.1/124585 1149013871 eng MIT theses are protected by copyright. They may be viewed, downloaded, or printed from this source but further reproduction or distribution in any format is prohibited without written permission. http://dspace.mit.edu/handle/1721.1/7582 29 pages application/pdf Massachusetts Institute of Technology |
spellingShingle | Sloan School of Management. Elkind, Daniel(Daniel Harris) A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title | A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title_full | A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title_fullStr | A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title_full_unstemmed | A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title_short | A reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
title_sort | reinforcement learning algorithm for efficient dynamic trading execution in the presence of signals |
topic | Sloan School of Management. |
url | https://hdl.handle.net/1721.1/124585 |
work_keys_str_mv | AT elkinddanieldanielharris areinforcementlearningalgorithmforefficientdynamictradingexecutioninthepresenceofsignals AT elkinddanieldanielharris reinforcementlearningalgorithmforefficientdynamictradingexecutioninthepresenceofsignals |