{"title":"一种加权平滑q -学习算法","authors":"V. Antony Vijesh;S. R. Shreyas","doi":"10.1109/LCSYS.2025.3551265","DOIUrl":null,"url":null,"abstract":"Q-learning and double Q-learning are well-known sample-based, off-policy reinforcement learning algorithms. However, Q-learning suffers from overestimation bias, while double Q-learning suffers from underestimation bias. To address these issues, this letter proposes a weighted smooth Q-learning (WSQL) algorithm. The proposed algorithm employs a weighted combination of the mellowmax operator and the log-sum-exp operator in place of the maximum operator. Firstly, a new stochastic approximation based result is derived and as a consequence the almost sure convergence of the proposed WSQL is presented. Further, a sufficient condition for the boundedness of WSQL algorithm is obtained. Numerical experiments are conducted on benchmark examples to validate the effectiveness of the proposed weighted smooth Q-learning algorithm.","PeriodicalId":37235,"journal":{"name":"IEEE Control Systems Letters","volume":"9 ","pages":"21-26"},"PeriodicalIF":2.4000,"publicationDate":"2025-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Weighted Smooth Q-Learning Algorithm\",\"authors\":\"V. Antony Vijesh;S. R. Shreyas\",\"doi\":\"10.1109/LCSYS.2025.3551265\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Q-learning and double Q-learning are well-known sample-based, off-policy reinforcement learning algorithms. However, Q-learning suffers from overestimation bias, while double Q-learning suffers from underestimation bias. To address these issues, this letter proposes a weighted smooth Q-learning (WSQL) algorithm. The proposed algorithm employs a weighted combination of the mellowmax operator and the log-sum-exp operator in place of the maximum operator. Firstly, a new stochastic approximation based result is derived and as a consequence the almost sure convergence of the proposed WSQL is presented. Further, a sufficient condition for the boundedness of WSQL algorithm is obtained. Numerical experiments are conducted on benchmark examples to validate the effectiveness of the proposed weighted smooth Q-learning algorithm.\",\"PeriodicalId\":37235,\"journal\":{\"name\":\"IEEE Control Systems Letters\",\"volume\":\"9 \",\"pages\":\"21-26\"},\"PeriodicalIF\":2.4000,\"publicationDate\":\"2025-03-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Control Systems Letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10925426/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Control Systems Letters","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10925426/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Q-learning and double Q-learning are well-known sample-based, off-policy reinforcement learning algorithms. However, Q-learning suffers from overestimation bias, while double Q-learning suffers from underestimation bias. To address these issues, this letter proposes a weighted smooth Q-learning (WSQL) algorithm. The proposed algorithm employs a weighted combination of the mellowmax operator and the log-sum-exp operator in place of the maximum operator. Firstly, a new stochastic approximation based result is derived and as a consequence the almost sure convergence of the proposed WSQL is presented. Further, a sufficient condition for the boundedness of WSQL algorithm is obtained. Numerical experiments are conducted on benchmark examples to validate the effectiveness of the proposed weighted smooth Q-learning algorithm.