将基于变压器的深度强化学习与 Black-Litterman 模型相结合,实现投资组合优化

Ruoyu SunXi'an Jiaotong-Liverpool University, School of Mathematics and Physics, Department of Financial and Actuarial Mathematics, Angelos StefanidisXi'an Jiaotong-Liverpool University Entrepreneur College, Zhengyong JiangXi'an Jiaotong-Liverpool University Entrepreneur College, Jionglong SuXi'an Jiaotong-Liverpool University Entrepreneur College
{"title":"将基于变压器的深度强化学习与 Black-Litterman 模型相结合,实现投资组合优化","authors":"Ruoyu SunXi'an Jiaotong-Liverpool University, School of Mathematics and Physics, Department of Financial and Actuarial Mathematics, Angelos StefanidisXi'an Jiaotong-Liverpool University Entrepreneur College, Zhengyong JiangXi'an Jiaotong-Liverpool University Entrepreneur College, Jionglong SuXi'an Jiaotong-Liverpool University Entrepreneur College","doi":"arxiv-2402.16609","DOIUrl":null,"url":null,"abstract":"As a model-free algorithm, deep reinforcement learning (DRL) agent learns and\nmakes decisions by interacting with the environment in an unsupervised way. In\nrecent years, DRL algorithms have been widely applied by scholars for portfolio\noptimization in consecutive trading periods, since the DRL agent can\ndynamically adapt to market changes and does not rely on the specification of\nthe joint dynamics across the assets. However, typical DRL agents for portfolio\noptimization cannot learn a policy that is aware of the dynamic correlation\nbetween portfolio asset returns. Since the dynamic correlations among portfolio\nassets are crucial in optimizing the portfolio, the lack of such knowledge\nmakes it difficult for the DRL agent to maximize the return per unit of risk,\nespecially when the target market permits short selling (i.e., the US stock\nmarket). In this research, we propose a hybrid portfolio optimization model\ncombining the DRL agent and the Black-Litterman (BL) model to enable the DRL\nagent to learn the dynamic correlation between the portfolio asset returns and\nimplement an efficacious long/short strategy based on the correlation.\nEssentially, the DRL agent is trained to learn the policy to apply the BL model\nto determine the target portfolio weights. To test our DRL agent, we construct\nthe portfolio based on all the Dow Jones Industrial Average constitute stocks.\nEmpirical results of the experiments conducted on real-world United States\nstock market data demonstrate that our DRL agent significantly outperforms\nvarious comparison portfolio choice strategies and alternative DRL frameworks\nby at least 42% in terms of accumulated return. In terms of the return per unit\nof risk, our DRL agent significantly outperforms various comparative portfolio\nchoice strategies and alternative strategies based on other machine learning\nframeworks.","PeriodicalId":501045,"journal":{"name":"arXiv - QuantFin - Portfolio Management","volume":"40 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Combining Transformer based Deep Reinforcement Learning with Black-Litterman Model for Portfolio Optimization\",\"authors\":\"Ruoyu SunXi'an Jiaotong-Liverpool University, School of Mathematics and Physics, Department of Financial and Actuarial Mathematics, Angelos StefanidisXi'an Jiaotong-Liverpool University Entrepreneur College, Zhengyong JiangXi'an Jiaotong-Liverpool University Entrepreneur College, Jionglong SuXi'an Jiaotong-Liverpool University Entrepreneur College\",\"doi\":\"arxiv-2402.16609\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As a model-free algorithm, deep reinforcement learning (DRL) agent learns and\\nmakes decisions by interacting with the environment in an unsupervised way. In\\nrecent years, DRL algorithms have been widely applied by scholars for portfolio\\noptimization in consecutive trading periods, since the DRL agent can\\ndynamically adapt to market changes and does not rely on the specification of\\nthe joint dynamics across the assets. However, typical DRL agents for portfolio\\noptimization cannot learn a policy that is aware of the dynamic correlation\\nbetween portfolio asset returns. Since the dynamic correlations among portfolio\\nassets are crucial in optimizing the portfolio, the lack of such knowledge\\nmakes it difficult for the DRL agent to maximize the return per unit of risk,\\nespecially when the target market permits short selling (i.e., the US stock\\nmarket). In this research, we propose a hybrid portfolio optimization model\\ncombining the DRL agent and the Black-Litterman (BL) model to enable the DRL\\nagent to learn the dynamic correlation between the portfolio asset returns and\\nimplement an efficacious long/short strategy based on the correlation.\\nEssentially, the DRL agent is trained to learn the policy to apply the BL model\\nto determine the target portfolio weights. To test our DRL agent, we construct\\nthe portfolio based on all the Dow Jones Industrial Average constitute stocks.\\nEmpirical results of the experiments conducted on real-world United States\\nstock market data demonstrate that our DRL agent significantly outperforms\\nvarious comparison portfolio choice strategies and alternative DRL frameworks\\nby at least 42% in terms of accumulated return. In terms of the return per unit\\nof risk, our DRL agent significantly outperforms various comparative portfolio\\nchoice strategies and alternative strategies based on other machine learning\\nframeworks.\",\"PeriodicalId\":501045,\"journal\":{\"name\":\"arXiv - QuantFin - Portfolio Management\",\"volume\":\"40 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Portfolio Management\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2402.16609\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Portfolio Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2402.16609","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

作为一种无模型算法,深度强化学习(DRL)代理通过与环境的无监督交互来学习和决策。近年来,DRL 算法被学者们广泛应用于连续交易期的投资组合优化,因为 DRL 代理可以动态地适应市场变化,并且不依赖于资产间联合动态的规范。然而,典型的用于投资组合优化的 DRL 代理无法学习到能够意识到投资组合资产收益之间动态相关性的策略。由于投资组合资产间的动态相关性对优化投资组合至关重要,缺乏这种知识使得 DRL 代理难以实现单位风险收益的最大化,尤其是在目标市场允许卖空的情况下(如美国股市)。在这项研究中,我们提出了一种混合投资组合优化模型,将 DRL 代理和布莱克-利特曼(Black-Litterman,BL)模型结合起来,使 DRL 代理能够学习投资组合资产收益之间的动态相关性,并根据相关性实施有效的多空策略。为了测试我们的 DRL 代理,我们以道琼斯工业平均指数构成的所有股票为基础构建了投资组合。在真实的美国股票市场数据上进行的实验结果表明,我们的 DRL 代理在累计收益方面明显优于各种比较投资组合选择策略和其他 DRL 框架,至少高出 42%。在单位风险收益方面,我们的 DRL 代理明显优于各种比较投资组合选择策略和基于其他机器学习框架的替代策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Combining Transformer based Deep Reinforcement Learning with Black-Litterman Model for Portfolio Optimization
As a model-free algorithm, deep reinforcement learning (DRL) agent learns and makes decisions by interacting with the environment in an unsupervised way. In recent years, DRL algorithms have been widely applied by scholars for portfolio optimization in consecutive trading periods, since the DRL agent can dynamically adapt to market changes and does not rely on the specification of the joint dynamics across the assets. However, typical DRL agents for portfolio optimization cannot learn a policy that is aware of the dynamic correlation between portfolio asset returns. Since the dynamic correlations among portfolio assets are crucial in optimizing the portfolio, the lack of such knowledge makes it difficult for the DRL agent to maximize the return per unit of risk, especially when the target market permits short selling (i.e., the US stock market). In this research, we propose a hybrid portfolio optimization model combining the DRL agent and the Black-Litterman (BL) model to enable the DRL agent to learn the dynamic correlation between the portfolio asset returns and implement an efficacious long/short strategy based on the correlation. Essentially, the DRL agent is trained to learn the policy to apply the BL model to determine the target portfolio weights. To test our DRL agent, we construct the portfolio based on all the Dow Jones Industrial Average constitute stocks. Empirical results of the experiments conducted on real-world United States stock market data demonstrate that our DRL agent significantly outperforms various comparison portfolio choice strategies and alternative DRL frameworks by at least 42% in terms of accumulated return. In terms of the return per unit of risk, our DRL agent significantly outperforms various comparative portfolio choice strategies and alternative strategies based on other machine learning frameworks.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信