An Online Home Energy Management System using Q-Learning and Deep Q-Learning

IF 3.8 3区 计算机科学 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
Hasan İzmitligil , Abdurrahman Karamancıoğlu
{"title":"An Online Home Energy Management System using Q-Learning and Deep Q-Learning","authors":"Hasan İzmitligil ,&nbsp;Abdurrahman Karamancıoğlu","doi":"10.1016/j.suscom.2024.101005","DOIUrl":null,"url":null,"abstract":"<div><p>The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.</p></div>","PeriodicalId":48686,"journal":{"name":"Sustainable Computing-Informatics & Systems","volume":"43 ","pages":"Article 101005"},"PeriodicalIF":3.8000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Sustainable Computing-Informatics & Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2210537924000507","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0

Abstract

The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.

使用 Q-Learning 和深度 Q-Learning 的在线家庭能源管理系统
由于通信技术和智能计量基础设施的进步,家庭能源管理系统的用户可以安排自己的实时能源消耗。本文提出了一种数据驱动策略,即在线家庭能源管理系统(ON-HEM),它使用强化学习算法(Q-Learning 和 Deep Q-Learning)来控制智能家居系统的最佳能耗。拟议的系统由电力资源(电网、光伏)、通信网络和电器组成,其代理分为四组:可延期、不可延期、功率水平可控和电动汽车。该系统降低了电费成本和高峰需求,同时考虑到了用户对真实数据不满的成本。我们使用 PyCharm 专业版软件对拟议的 ON-HEM 进行了仿真,考虑了不同的定价方法(实时定价和使用时间定价)以及 Q-Learning 和 Deep Q-Learning (DQL) 算法。研究结果表明,DQL 比 Q-Learning 更优越,而且建议的 ON-HEM 在降低高峰需求、电费和客户不满成本方面也很有效。通过使用 IBM SPSS 统计软件将基于模拟的研究结果与实际数据相结合,验证了所建议系统的效率和可靠性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Sustainable Computing-Informatics & Systems
Sustainable Computing-Informatics & Systems COMPUTER SCIENCE, HARDWARE & ARCHITECTUREC-COMPUTER SCIENCE, INFORMATION SYSTEMS
CiteScore
10.70
自引率
4.40%
发文量
142
期刊介绍: Sustainable computing is a rapidly expanding research area spanning the fields of computer science and engineering, electrical engineering as well as other engineering disciplines. The aim of Sustainable Computing: Informatics and Systems (SUSCOM) is to publish the myriad research findings related to energy-aware and thermal-aware management of computing resource. Equally important is a spectrum of related research issues such as applications of computing that can have ecological and societal impacts. SUSCOM publishes original and timely research papers and survey articles in current areas of power, energy, temperature, and environment related research areas of current importance to readers. SUSCOM has an editorial board comprising prominent researchers from around the world and selects competitively evaluated peer-reviewed papers.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信