D. S. Keerthi, P. Vishwanath, Kothuri Parashu Ramulu, Gopinath Anjinappa, Hirald Dwaraka Praveena
{"title":"基于大时空图变压器模型的无线网络用户关联与功率分配联合优化","authors":"D. S. Keerthi, P. Vishwanath, Kothuri Parashu Ramulu, Gopinath Anjinappa, Hirald Dwaraka Praveena","doi":"10.1002/itl2.70131","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>In this era, Wireless Communication Networks (WCNs) need dynamic and adaptive resource allocation approaches to handle user association and power allocation specifically under multi-connectivity and diverse traffic conditions. However, the conventional approaches struggle due to high computational cost, poor adaptability, and limited generalization. Therefore, this research proposes a large Spatio-Temporal Graph Transformer-based Reinforcement Learning (STGT-RL) model to jointly optimize user association and power allocation in large-scale WCNs. Initially, the network topology is designed using graph representations and incorporates a hybrid encoder that integrates Graph Transformers for spatial user-Base Station (BS) relationships and Spatio-Temporal Transformers for capturing time-varying traffic and channel states. Further, to ensure adaptive decision-making, a Transformer-RL policy agent is trained through a multi-objective reward function that assists in balancing throughput maximization and power efficiency. Furthermore, to enable stable policy learning, the model is initially trained using high-quality supervision from CRFSMA-generated labels, followed by reinforcement-based policy refinement. Hence, the experimental results are simulated on WCN environments to demonstrate that the proposed STGT-RL significantly outperforms baseline deep learning and heuristic-based methods in terms of throughput, energy efficiency, and fairness.</p>\n </div>","PeriodicalId":100725,"journal":{"name":"Internet Technology Letters","volume":"8 6","pages":""},"PeriodicalIF":0.5000,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Joint Optimization of User Association and Power Allocation in Wireless Networks Using a Large Spatio-Temporal Graph Transformer Model\",\"authors\":\"D. S. Keerthi, P. Vishwanath, Kothuri Parashu Ramulu, Gopinath Anjinappa, Hirald Dwaraka Praveena\",\"doi\":\"10.1002/itl2.70131\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>In this era, Wireless Communication Networks (WCNs) need dynamic and adaptive resource allocation approaches to handle user association and power allocation specifically under multi-connectivity and diverse traffic conditions. However, the conventional approaches struggle due to high computational cost, poor adaptability, and limited generalization. Therefore, this research proposes a large Spatio-Temporal Graph Transformer-based Reinforcement Learning (STGT-RL) model to jointly optimize user association and power allocation in large-scale WCNs. Initially, the network topology is designed using graph representations and incorporates a hybrid encoder that integrates Graph Transformers for spatial user-Base Station (BS) relationships and Spatio-Temporal Transformers for capturing time-varying traffic and channel states. Further, to ensure adaptive decision-making, a Transformer-RL policy agent is trained through a multi-objective reward function that assists in balancing throughput maximization and power efficiency. Furthermore, to enable stable policy learning, the model is initially trained using high-quality supervision from CRFSMA-generated labels, followed by reinforcement-based policy refinement. Hence, the experimental results are simulated on WCN environments to demonstrate that the proposed STGT-RL significantly outperforms baseline deep learning and heuristic-based methods in terms of throughput, energy efficiency, and fairness.</p>\\n </div>\",\"PeriodicalId\":100725,\"journal\":{\"name\":\"Internet Technology Letters\",\"volume\":\"8 6\",\"pages\":\"\"},\"PeriodicalIF\":0.5000,\"publicationDate\":\"2025-09-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Internet Technology Letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/itl2.70131\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"TELECOMMUNICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet Technology Letters","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/itl2.70131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
Joint Optimization of User Association and Power Allocation in Wireless Networks Using a Large Spatio-Temporal Graph Transformer Model
In this era, Wireless Communication Networks (WCNs) need dynamic and adaptive resource allocation approaches to handle user association and power allocation specifically under multi-connectivity and diverse traffic conditions. However, the conventional approaches struggle due to high computational cost, poor adaptability, and limited generalization. Therefore, this research proposes a large Spatio-Temporal Graph Transformer-based Reinforcement Learning (STGT-RL) model to jointly optimize user association and power allocation in large-scale WCNs. Initially, the network topology is designed using graph representations and incorporates a hybrid encoder that integrates Graph Transformers for spatial user-Base Station (BS) relationships and Spatio-Temporal Transformers for capturing time-varying traffic and channel states. Further, to ensure adaptive decision-making, a Transformer-RL policy agent is trained through a multi-objective reward function that assists in balancing throughput maximization and power efficiency. Furthermore, to enable stable policy learning, the model is initially trained using high-quality supervision from CRFSMA-generated labels, followed by reinforcement-based policy refinement. Hence, the experimental results are simulated on WCN environments to demonstrate that the proposed STGT-RL significantly outperforms baseline deep learning and heuristic-based methods in terms of throughput, energy efficiency, and fairness.