{"title":"强化学习的溢价控制","authors":"L. Palmborg, F. Lindskog","doi":"10.1017/asb.2023.13","DOIUrl":null,"url":null,"abstract":"Abstract We consider a premium control problem in discrete time, formulated in terms of a Markov decision process. In a simplified setting, the optimal premium rule can be derived with dynamic programming methods. However, these classical methods are not feasible in a more realistic setting due to the dimension of the state space and lack of explicit expressions for transition probabilities. We explore reinforcement learning techniques, using function approximation, to solve the premium control problem for realistic stochastic models. We illustrate the appropriateness of the approximate optimal premium rule compared with the true optimal premium rule in a simplified setting and further demonstrate that the approximate optimal premium rule outperforms benchmark rules in more realistic settings where classical approaches fail.","PeriodicalId":8617,"journal":{"name":"ASTIN Bulletin","volume":"83 1","pages":"233 - 257"},"PeriodicalIF":1.7000,"publicationDate":"2023-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Premium control with reinforcement learning\",\"authors\":\"L. Palmborg, F. Lindskog\",\"doi\":\"10.1017/asb.2023.13\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Abstract We consider a premium control problem in discrete time, formulated in terms of a Markov decision process. In a simplified setting, the optimal premium rule can be derived with dynamic programming methods. However, these classical methods are not feasible in a more realistic setting due to the dimension of the state space and lack of explicit expressions for transition probabilities. We explore reinforcement learning techniques, using function approximation, to solve the premium control problem for realistic stochastic models. We illustrate the appropriateness of the approximate optimal premium rule compared with the true optimal premium rule in a simplified setting and further demonstrate that the approximate optimal premium rule outperforms benchmark rules in more realistic settings where classical approaches fail.\",\"PeriodicalId\":8617,\"journal\":{\"name\":\"ASTIN Bulletin\",\"volume\":\"83 1\",\"pages\":\"233 - 257\"},\"PeriodicalIF\":1.7000,\"publicationDate\":\"2023-04-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ASTIN Bulletin\",\"FirstCategoryId\":\"96\",\"ListUrlMain\":\"https://doi.org/10.1017/asb.2023.13\",\"RegionNum\":3,\"RegionCategory\":\"经济学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ECONOMICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ASTIN Bulletin","FirstCategoryId":"96","ListUrlMain":"https://doi.org/10.1017/asb.2023.13","RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ECONOMICS","Score":null,"Total":0}
Abstract We consider a premium control problem in discrete time, formulated in terms of a Markov decision process. In a simplified setting, the optimal premium rule can be derived with dynamic programming methods. However, these classical methods are not feasible in a more realistic setting due to the dimension of the state space and lack of explicit expressions for transition probabilities. We explore reinforcement learning techniques, using function approximation, to solve the premium control problem for realistic stochastic models. We illustrate the appropriateness of the approximate optimal premium rule compared with the true optimal premium rule in a simplified setting and further demonstrate that the approximate optimal premium rule outperforms benchmark rules in more realistic settings where classical approaches fail.
期刊介绍:
ASTIN Bulletin publishes papers that are relevant to any branch of actuarial science and insurance mathematics. Its papers are quantitative and scientific in nature, and draw on theory and methods developed in any branch of the mathematical sciences including actuarial mathematics, statistics, probability, financial mathematics and econometrics.