Multi-Agent Q-Learning for Competitive Spectrum Access in Cognitive Radio Systems

Husheng Li
{"title":"Multi-Agent Q-Learning for Competitive Spectrum Access in Cognitive Radio Systems","authors":"Husheng Li","doi":"10.1109/SDR.2010.5507919","DOIUrl":null,"url":null,"abstract":"Resource allocation is an important issue in cognitive radio systems. It can be done by carrying out negotiation among secondary users. However, significant overhead may be incurred by the negotiation since the negotiation needs to be done frequently due to the rapid change of primary users' activity. In this paper, an Aloha-like spectrum access scheme without negotiation is considered for multi-user and multi-channel cognitive radio systems. To avoid collision incurred by the lack of coordination, each secondary user learns how to select channels according to its experience. Multi-agent reinforcement leaning (MARL) is applied in the framework of $Q$-learning by considering other secondary users as a part of the environment. A rigorous proof of the convergence of $Q$-learning is provided via the similarity between the $Q$-learning and Robinson-Monro algorithm, as well as the analysis of the corresponding ordinary differential equation (via Lyapunov function). The performance of learning (speed and gain in utility) is evaluated by numerical simulations.","PeriodicalId":236424,"journal":{"name":"2010 Fifth IEEE Workshop on Networking Technologies for Software Defined Radio Networks (SDR)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"48","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Fifth IEEE Workshop on Networking Technologies for Software Defined Radio Networks (SDR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SDR.2010.5507919","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 48

Abstract

Resource allocation is an important issue in cognitive radio systems. It can be done by carrying out negotiation among secondary users. However, significant overhead may be incurred by the negotiation since the negotiation needs to be done frequently due to the rapid change of primary users' activity. In this paper, an Aloha-like spectrum access scheme without negotiation is considered for multi-user and multi-channel cognitive radio systems. To avoid collision incurred by the lack of coordination, each secondary user learns how to select channels according to its experience. Multi-agent reinforcement leaning (MARL) is applied in the framework of $Q$-learning by considering other secondary users as a part of the environment. A rigorous proof of the convergence of $Q$-learning is provided via the similarity between the $Q$-learning and Robinson-Monro algorithm, as well as the analysis of the corresponding ordinary differential equation (via Lyapunov function). The performance of learning (speed and gain in utility) is evaluated by numerical simulations.
认知无线电系统中竞争频谱接入的多智能体q学习
资源分配是认知无线电系统中的一个重要问题。这可以通过在二级用户之间进行协商来实现。但是,由于主要用户活动的快速变化,需要频繁地进行协商,因此协商可能会产生很大的开销。针对多用户多信道认知无线电系统,研究了一种无协商的类aloha频谱接入方案。为了避免由于缺乏协调而产生的冲突,每个二级用户根据自己的经验学习如何选择信道。通过将其他辅助用户视为环境的一部分,将多智能体强化学习(MARL)应用于$Q$学习框架中。通过$Q$学习与Robinson-Monro算法之间的相似性,以及对相应的常微分方程(通过Lyapunov函数)的分析,提供了$Q$学习收敛性的严格证明。通过数值模拟对学习性能(速度和效用增益)进行了评价。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信