{"title":"Optimizing Deep Reinforcement Learning for American Put Option Hedging","authors":"Reilly Pickard, F. Wredenhagen, Y. Lawryshyn","doi":"arxiv-2405.08602","DOIUrl":null,"url":null,"abstract":"This paper contributes to the existing literature on hedging American options\nwith Deep Reinforcement Learning (DRL). The study first investigates\nhyperparameter impact on hedging performance, considering learning rates,\ntraining episodes, neural network architectures, training steps, and\ntransaction cost penalty functions. Results highlight the importance of\navoiding certain combinations, such as high learning rates with a high number\nof training episodes or low learning rates with few training episodes and\nemphasize the significance of utilizing moderate values for optimal outcomes.\nAdditionally, the paper warns against excessive training steps to prevent\ninstability and demonstrates the superiority of a quadratic transaction cost\npenalty function over a linear version. This study then expands upon the work\nof Pickard et al. (2024), who utilize a Chebyshev interpolation option pricing\nmethod to train DRL agents with market calibrated stochastic volatility models.\nWhile the results of Pickard et al. (2024) showed that these DRL agents achieve\nsatisfactory performance on empirical asset paths, this study introduces a\nnovel approach where new agents at weekly intervals to newly calibrated\nstochastic volatility models. Results show DRL agents re-trained using weekly\nmarket data surpass the performance of those trained solely on the sale date.\nFurthermore, the paper demonstrates that both single-train and weekly-train DRL\nagents outperform the Black-Scholes Delta method at transaction costs of 1% and\n3%. This practical relevance suggests that practitioners can leverage readily\navailable market data to train DRL agents for effective hedging of options in\ntheir portfolios.","PeriodicalId":501128,"journal":{"name":"arXiv - QuantFin - Risk Management","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Risk Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.08602","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper contributes to the existing literature on hedging American options
with Deep Reinforcement Learning (DRL). The study first investigates
hyperparameter impact on hedging performance, considering learning rates,
training episodes, neural network architectures, training steps, and
transaction cost penalty functions. Results highlight the importance of
avoiding certain combinations, such as high learning rates with a high number
of training episodes or low learning rates with few training episodes and
emphasize the significance of utilizing moderate values for optimal outcomes.
Additionally, the paper warns against excessive training steps to prevent
instability and demonstrates the superiority of a quadratic transaction cost
penalty function over a linear version. This study then expands upon the work
of Pickard et al. (2024), who utilize a Chebyshev interpolation option pricing
method to train DRL agents with market calibrated stochastic volatility models.
While the results of Pickard et al. (2024) showed that these DRL agents achieve
satisfactory performance on empirical asset paths, this study introduces a
novel approach where new agents at weekly intervals to newly calibrated
stochastic volatility models. Results show DRL agents re-trained using weekly
market data surpass the performance of those trained solely on the sale date.
Furthermore, the paper demonstrates that both single-train and weekly-train DRL
agents outperform the Black-Scholes Delta method at transaction costs of 1% and
3%. This practical relevance suggests that practitioners can leverage readily
available market data to train DRL agents for effective hedging of options in
their portfolios.