动态预期风险下期权定价与套期保值的深度强化学习

IF 1.5 4区 经济学 Q3 BUSINESS, FINANCE
Option Pricing, Saeed Marzban, E. Delage, Jonathan Yu-Meng Li
{"title":"动态预期风险下期权定价与套期保值的深度强化学习","authors":"Option Pricing, Saeed Marzban, E. Delage, Jonathan Yu-Meng Li","doi":"10.1080/14697688.2023.2244531","DOIUrl":null,"url":null,"abstract":"Recently equal risk pricing, a framework for fair derivative pricing, was extended to consider dynamic risk measures. However, all current implementations either employ a static risk measure that violates time consistency, or are based on traditional dynamic programing solution schemes that are impracticable in problems with a large number of underlying assets (due to the curse of dimensionality) or with incomplete asset dynamics information. In this paper, we extend for the first time a famous off-policy deterministic actor-critic deep reinforcement learning (ACRL) algorithm to the problem of solving a risk averse Markov decision process that models risk using a time consistent recursive expectile risk measure. This new ACRL algorithm allows us to identify high quality time consistent hedging policies (and equal risk prices) for options, such as basket options, that cannot be handled using traditional methods, or in context where only historical trajectories of the underlying assets are available. Our numerical experiments, which involve both a simple vanilla option and a more exotic basket option, confirm that the new ACRL algorithm can produce (1) in simple environments, nearly optimal hedging policies, and highly accurate prices, simultaneously for a range of maturities (2) in complex environments, good quality policies and prices using reasonable amount of computing resources; and (3) overall, hedging strategies that actually outperform the strategies produced using static risk measures when the risk is evaluated at later points of time.","PeriodicalId":20747,"journal":{"name":"Quantitative Finance","volume":"81 1","pages":"1411 - 1430"},"PeriodicalIF":1.5000,"publicationDate":"2023-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep reinforcement learning for option pricing and hedging under dynamic expectile risk measures\",\"authors\":\"Option Pricing, Saeed Marzban, E. Delage, Jonathan Yu-Meng Li\",\"doi\":\"10.1080/14697688.2023.2244531\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recently equal risk pricing, a framework for fair derivative pricing, was extended to consider dynamic risk measures. However, all current implementations either employ a static risk measure that violates time consistency, or are based on traditional dynamic programing solution schemes that are impracticable in problems with a large number of underlying assets (due to the curse of dimensionality) or with incomplete asset dynamics information. In this paper, we extend for the first time a famous off-policy deterministic actor-critic deep reinforcement learning (ACRL) algorithm to the problem of solving a risk averse Markov decision process that models risk using a time consistent recursive expectile risk measure. This new ACRL algorithm allows us to identify high quality time consistent hedging policies (and equal risk prices) for options, such as basket options, that cannot be handled using traditional methods, or in context where only historical trajectories of the underlying assets are available. Our numerical experiments, which involve both a simple vanilla option and a more exotic basket option, confirm that the new ACRL algorithm can produce (1) in simple environments, nearly optimal hedging policies, and highly accurate prices, simultaneously for a range of maturities (2) in complex environments, good quality policies and prices using reasonable amount of computing resources; and (3) overall, hedging strategies that actually outperform the strategies produced using static risk measures when the risk is evaluated at later points of time.\",\"PeriodicalId\":20747,\"journal\":{\"name\":\"Quantitative Finance\",\"volume\":\"81 1\",\"pages\":\"1411 - 1430\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2023-09-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Quantitative Finance\",\"FirstCategoryId\":\"96\",\"ListUrlMain\":\"https://doi.org/10.1080/14697688.2023.2244531\",\"RegionNum\":4,\"RegionCategory\":\"经济学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"BUSINESS, FINANCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quantitative Finance","FirstCategoryId":"96","ListUrlMain":"https://doi.org/10.1080/14697688.2023.2244531","RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"BUSINESS, FINANCE","Score":null,"Total":0}
引用次数: 0

摘要

最近,公允衍生品定价框架等风险定价被扩展到考虑动态风险度量。然而,所有当前的实现要么采用违反时间一致性的静态风险度量,要么基于传统的动态规划解决方案,这些方案在具有大量底层资产(由于维度的诅咒)或资产动态信息不完整的问题中是不切实际的。在本文中,我们首次将一个著名的off-policy deterministic actor-critic深度强化学习(ACRL)算法扩展到解决一个风险厌恶的马尔可夫决策过程的问题,该决策过程使用时间一致的递归预期风险度量来建模风险。这种新的ACRL算法使我们能够为期权(如一揽子期权)确定高质量的时间一致的对冲政策(和相等的风险价格),这些期权无法使用传统方法处理,或者只能在基础资产的历史轨迹可用的情况下处理。我们的数值实验,包括一个简单的香草期权和一个更奇特的篮子期权,证实了新的ACRL算法可以产生(1)在简单的环境中,几乎最优的对冲政策和高度准确的价格,同时对于一系列期限(2)在复杂的环境中,良好的质量政策和价格使用合理的计算资源;(3)总体而言,当风险在稍后的时间点进行评估时,对冲策略实际上优于使用静态风险度量产生的策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Deep reinforcement learning for option pricing and hedging under dynamic expectile risk measures
Recently equal risk pricing, a framework for fair derivative pricing, was extended to consider dynamic risk measures. However, all current implementations either employ a static risk measure that violates time consistency, or are based on traditional dynamic programing solution schemes that are impracticable in problems with a large number of underlying assets (due to the curse of dimensionality) or with incomplete asset dynamics information. In this paper, we extend for the first time a famous off-policy deterministic actor-critic deep reinforcement learning (ACRL) algorithm to the problem of solving a risk averse Markov decision process that models risk using a time consistent recursive expectile risk measure. This new ACRL algorithm allows us to identify high quality time consistent hedging policies (and equal risk prices) for options, such as basket options, that cannot be handled using traditional methods, or in context where only historical trajectories of the underlying assets are available. Our numerical experiments, which involve both a simple vanilla option and a more exotic basket option, confirm that the new ACRL algorithm can produce (1) in simple environments, nearly optimal hedging policies, and highly accurate prices, simultaneously for a range of maturities (2) in complex environments, good quality policies and prices using reasonable amount of computing resources; and (3) overall, hedging strategies that actually outperform the strategies produced using static risk measures when the risk is evaluated at later points of time.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Quantitative Finance
Quantitative Finance 社会科学-数学跨学科应用
CiteScore
3.20
自引率
7.70%
发文量
102
审稿时长
4-8 weeks
期刊介绍: The frontiers of finance are shifting rapidly, driven in part by the increasing use of quantitative methods in the field. Quantitative Finance welcomes original research articles that reflect the dynamism of this area. The journal provides an interdisciplinary forum for presenting both theoretical and empirical approaches and offers rapid publication of original new work with high standards of quality. The readership is broad, embracing researchers and practitioners across a range of specialisms and within a variety of organizations. All articles should aim to be of interest to this broad readership.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信