Double Deep Q-Learning for Optimal Execution

Q3 Mathematics
Brian Ning, Franco Ho Ting Ling, S. Jaimungal
{"title":"Double Deep Q-Learning for Optimal Execution","authors":"Brian Ning, Franco Ho Ting Ling, S. Jaimungal","doi":"10.1080/1350486X.2022.2077783","DOIUrl":null,"url":null,"abstract":"ABSTRACT Optimal trade execution is an important problem faced by essentially all traders. Much research into optimal execution uses stringent model assumptions and applies continuous time stochastic control to solve them. Here, we instead take a model free approach and develop a variation of Deep Q-Learning to estimate the optimal actions of a trader. The model is a fully connected Neural Network trained using Experience Replay and Double DQN with input features given by the current state of the limit order book, other trading signals, and available execution actions, while the output is the Q-value function estimating the future rewards under an arbitrary action. We apply our model to nine different stocks and find that it outperforms the standard benchmark approach on most stocks using the measures of (i) mean and median out-performance, (ii) probability of out-performance, and (iii) gain-loss ratios.","PeriodicalId":35818,"journal":{"name":"Applied Mathematical Finance","volume":"1 1","pages":"361 - 380"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"52","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Mathematical Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/1350486X.2022.2077783","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Mathematics","Score":null,"Total":0}
引用次数: 52

Abstract

ABSTRACT Optimal trade execution is an important problem faced by essentially all traders. Much research into optimal execution uses stringent model assumptions and applies continuous time stochastic control to solve them. Here, we instead take a model free approach and develop a variation of Deep Q-Learning to estimate the optimal actions of a trader. The model is a fully connected Neural Network trained using Experience Replay and Double DQN with input features given by the current state of the limit order book, other trading signals, and available execution actions, while the output is the Q-value function estimating the future rewards under an arbitrary action. We apply our model to nine different stocks and find that it outperforms the standard benchmark approach on most stocks using the measures of (i) mean and median out-performance, (ii) probability of out-performance, and (iii) gain-loss ratios.
双深度q -学习优化执行
最优交易执行基本上是所有交易者都面临的一个重要问题。许多关于最优执行的研究都采用严格的模型假设,并采用连续时间随机控制来解决这些问题。在这里,我们采用了一种无模型的方法,并开发了一种深度q学习的变体来估计交易者的最佳行为。该模型是一个使用Experience Replay和Double DQN训练的全连接神经网络,其输入特征由限价单的当前状态、其他交易信号和可用的执行动作给出,而输出是估计任意动作下未来奖励的q值函数。我们将我们的模型应用于9只不同的股票,发现它在大多数股票上的表现优于标准基准方法,使用(i)平均和中位数表现,(ii)表现优异的概率,以及(iii)损益比。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Applied Mathematical Finance
Applied Mathematical Finance Economics, Econometrics and Finance-Finance
CiteScore
2.30
自引率
0.00%
发文量
6
期刊介绍: The journal encourages the confident use of applied mathematics and mathematical modelling in finance. The journal publishes papers on the following: •modelling of financial and economic primitives (interest rates, asset prices etc); •modelling market behaviour; •modelling market imperfections; •pricing of financial derivative securities; •hedging strategies; •numerical methods; •financial engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信