{"title":"基于深度强化学习的物联网移动数据收集节能和情境感知轨迹规划","authors":"Sana Benhamaid, Hicham Lakhlef, A. Bouabdallah","doi":"10.23919/softcom55329.2022.9911304","DOIUrl":null,"url":null,"abstract":"IoT networks are often composed of spatially distributed nodes. This is why mobile data collection (MDC) emerged as an efficient solution to gather data from IoT networks that tolerate delay. In this paper, we study the use of reinforcement learning (RL) to plan the data collection trajectory of a mobile node (MN) in cluster-based IoT networks. Most of the existing solutions use static methods. However, in a context where the MN has little information (no previous data set) about the environment and where the environment is subject to changes (cluster mobility, etc.), we want the MN to learn an energy-efficient trajectory and adapt the trajectory to the significant changes in the environment. For that purpose, we will train two reinforcement learning (RL) algorithms: Q-learning and state-action-reward-state-action (SARSA) combined with deep learning (DL). This solution will allow us to maximize the collected data while minimizing the energy consumption of the MN. These algorithms will also adapt the trajectory of the MN to the signiflcant changes in the environment.","PeriodicalId":261625,"journal":{"name":"2022 International Conference on Software, Telecommunications and Computer Networks (SoftCOM)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Energy-Efficient and Context-aware Trajectory Planning for Mobile Data Collection in IoT using Deep Reinforcement Learning\",\"authors\":\"Sana Benhamaid, Hicham Lakhlef, A. Bouabdallah\",\"doi\":\"10.23919/softcom55329.2022.9911304\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"IoT networks are often composed of spatially distributed nodes. This is why mobile data collection (MDC) emerged as an efficient solution to gather data from IoT networks that tolerate delay. In this paper, we study the use of reinforcement learning (RL) to plan the data collection trajectory of a mobile node (MN) in cluster-based IoT networks. Most of the existing solutions use static methods. However, in a context where the MN has little information (no previous data set) about the environment and where the environment is subject to changes (cluster mobility, etc.), we want the MN to learn an energy-efficient trajectory and adapt the trajectory to the significant changes in the environment. For that purpose, we will train two reinforcement learning (RL) algorithms: Q-learning and state-action-reward-state-action (SARSA) combined with deep learning (DL). This solution will allow us to maximize the collected data while minimizing the energy consumption of the MN. These algorithms will also adapt the trajectory of the MN to the signiflcant changes in the environment.\",\"PeriodicalId\":261625,\"journal\":{\"name\":\"2022 International Conference on Software, Telecommunications and Computer Networks (SoftCOM)\",\"volume\":\"47 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-09-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 International Conference on Software, Telecommunications and Computer Networks (SoftCOM)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/softcom55329.2022.9911304\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Software, Telecommunications and Computer Networks (SoftCOM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/softcom55329.2022.9911304","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Energy-Efficient and Context-aware Trajectory Planning for Mobile Data Collection in IoT using Deep Reinforcement Learning
IoT networks are often composed of spatially distributed nodes. This is why mobile data collection (MDC) emerged as an efficient solution to gather data from IoT networks that tolerate delay. In this paper, we study the use of reinforcement learning (RL) to plan the data collection trajectory of a mobile node (MN) in cluster-based IoT networks. Most of the existing solutions use static methods. However, in a context where the MN has little information (no previous data set) about the environment and where the environment is subject to changes (cluster mobility, etc.), we want the MN to learn an energy-efficient trajectory and adapt the trajectory to the significant changes in the environment. For that purpose, we will train two reinforcement learning (RL) algorithms: Q-learning and state-action-reward-state-action (SARSA) combined with deep learning (DL). This solution will allow us to maximize the collected data while minimizing the energy consumption of the MN. These algorithms will also adapt the trajectory of the MN to the signiflcant changes in the environment.