{"title":"Topology Design for Data Center Networks Using Deep Reinforcement Learning","authors":"Haoran Qi, Zhan Shu, Xiaomin Chen","doi":"10.1109/ICOIN56518.2023.10048955","DOIUrl":null,"url":null,"abstract":"This paper is concerned with the topology design of data center networks (DCNs) for low latency and fewer links using deep reinforcement learning (DRL). Starting from a K-vertex-connected graph, we propose an interactive framework with single-objective and multi-objective DRL agents to learn DCN topologies for given node traffic matrices by choosing link matrices to represent the states and actions as well as using the average shortest path length together with action penalty terms as reward feedback. Comparisons with commonly used DCN topologies are given to show the effectiveness and merits of our method. The results reveal that our learned topologies could achieve lower delay compared with common DCN topologies. Moreover, we believe that the method can be extended to other topology metrics, e.g., throughput, by simply modifying the reward functions.","PeriodicalId":285763,"journal":{"name":"2023 International Conference on Information Networking (ICOIN)","volume":"123 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Information Networking (ICOIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICOIN56518.2023.10048955","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper is concerned with the topology design of data center networks (DCNs) for low latency and fewer links using deep reinforcement learning (DRL). Starting from a K-vertex-connected graph, we propose an interactive framework with single-objective and multi-objective DRL agents to learn DCN topologies for given node traffic matrices by choosing link matrices to represent the states and actions as well as using the average shortest path length together with action penalty terms as reward feedback. Comparisons with commonly used DCN topologies are given to show the effectiveness and merits of our method. The results reveal that our learned topologies could achieve lower delay compared with common DCN topologies. Moreover, we believe that the method can be extended to other topology metrics, e.g., throughput, by simply modifying the reward functions.