Full text

Turn on search term navigation

© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

The integration of renewable energy sources, such as rooftop solar panels, into smart grids poses significant challenges for managing customer-side battery storage. In response, this paper introduces a novel reinforcement learning (RL) approach aimed at optimizing the coordination of these batteries. Our approach utilizes a single-agent, multi-environment RL system designed to balance power saving, customer satisfaction, and fairness in power distribution. The RL agent dynamically allocates charging power while accounting for individual battery levels and grid constraints, employing an actor–critic algorithm. The actor determines the optimal charging power based on real-time conditions, while the critic iteratively refines the policy to enhance overall performance. The key advantages of our approach include: (1) Adaptive Power Allocation: The RL agent effectively reduces overall power consumption by optimizing grid power allocation, leading to more efficient energy use. (2) Enhanced Customer Satisfaction: By increasing the total available power from the grid, our approach significantly reduces instances of battery levels falling below the critical state of charge (SoC), thereby improving customer satisfaction. (3) Fair Power Distribution: Fairness improvements are notable, with the highest fair reward rising by 173.7% across different scenarios, demonstrating the effectiveness of our method in minimizing discrepancies in power distribution. (4) Improved Total Reward: The total reward also shows a significant increase, up by 94.1%, highlighting the efficiency of our RL-based approach. Experimental results using a real-world dataset confirm that our RL approach markedly improves fairness, power efficiency, and customer satisfaction, underscoring its potential for optimizing smart grid operations and energy management systems.

Details

Title
Reinforcement Learning for Fair and Efficient Charging Coordination for Smart Grid
Author
Elshazly, Amr A 1   VIAFID ORCID Logo  ; Badr, Mahmoud M 2   VIAFID ORCID Logo  ; Mahmoud, Mohamed 3   VIAFID ORCID Logo  ; Eberle, William 1 ; Alsabaan, Maazen 4   VIAFID ORCID Logo  ; Ibrahem, Mohamed I 5   VIAFID ORCID Logo 

 Department of Computer Science, Tennessee Technological University, Cookeville, TN 38505, USA; [email protected] (A.A.E.); [email protected] (W.E.) 
 Department of Network and Computer Security, College of Engineering, SUNY Polytechnic Institute, Utica, NY 13502, USA; [email protected]; Department of Electrical Engineering, Faculty of Engineering at Shoubra, Benha University, Cairo 11672, Egypt; [email protected] 
 Department of Electrical and Computer Engineering, Tennessee Technological University, Cookeville, TN 38505, USA 
 Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi Arabia; [email protected] 
 Department of Electrical Engineering, Faculty of Engineering at Shoubra, Benha University, Cairo 11672, Egypt; [email protected]; School of Computer and Cyber Sciences, Augusta University, Augusta, GA 30912, USA 
First page
4557
Publication year
2024
Publication date
2024
Publisher
MDPI AG
e-ISSN
19961073
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3110464522
Copyright
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.