Next Article in Journal
Thermomechanical and Thermofluid-Dynamic Coupled Analysis of the Top Cap Region of the Water-Cooled Lithium Lead Breeding Blanket for the EU DEMO Fusion Reactor
Next Article in Special Issue
A Diagnosis Method of Power Flow Convergence Failure for Bulk Power Systems Based on Intermediate Iteration Data
Previous Article in Journal
Design and Optimization of Photovoltaic System in Full-Chain Ground-Based Validation System of Space Solar Power Station
Previous Article in Special Issue
Ground Fault Detection Based on Fault Data Stitching and Image Generation of Resonant Grounding Distribution Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Microgrid Collaborative Optimization Scheduling Using an Improved Multi-Agent Soft Actor-Critic Algorithm

1
School of Electrical Engineering, Northeast Electric Power University, Jilin 132012, China
2
State Grid Jining Power Supply Company, Jining 272000, China
*
Author to whom correspondence should be addressed.
Submission received: 26 February 2023 / Revised: 1 April 2023 / Accepted: 3 April 2023 / Published: 5 April 2023
(This article belongs to the Special Issue Modern Power System Stability and Optimal Operating)

Abstract

:
The implementation of a multi-microgrid (MMG) system with multiple renewable energy sources enables the facilitation of electricity trading. To tackle the energy management problem of an MMG system, which consists of multiple renewable energy microgrids belonging to different operating entities, this paper proposes an MMG collaborative optimization scheduling model based on a multi-agent centralized training distributed execution framework. To enhance the generalization ability of dealing with various uncertainties, we also propose an improved multi-agent soft actor-critic (MASAC) algorithm, which facilitates energy transactions between multi-agents in MMG, and employs automated machine learning (AutoML) to optimize the MASAC hyperparameters to further improve the generalization of deep reinforcement learning (DRL). The test results demonstrate that the proposed method successfully achieves power complementarity between different entities and reduces the MMG system’s operating cost. Additionally, the proposal significantly outperforms other state-of-the-art reinforcement learning algorithms with better economy and higher calculation efficiency.

1. Introduction

To achieve sustainable social development, it is imperative to embrace clean, low-carbon, and sustainable energy sources [1,2]. However, due to the inherent uncertainty of renewable energy, the integration of multiple renewable energy sources in the form of microgrid (MG) has played a significant role in promoting the consumption of renewable energy [3,4,5]. As technology advances, connecting multiple microgrids (MGs) within the same power distribution area can unlock the potential of various flexible resources, enabling the complementary utilization of multi-microgrid (MMG) energy [6]. In addition, this approach further promotes the consumption of various renewable energy sources, which has emerged as a new trend in development [7,8]. However, the energy interaction between multiple MGs involves complex transaction relationships, leading to significant challenges in system regulation. In this case, it is of great significance to investigate the collaborative optimal dispatch of MMG with electric energy interaction to fully exploit the potential of renewable energy sources and ensure efficient system regulation.
Existing research has made significant progress in addressing the complexity of managing MMG energy. Ref. [9] proposes optimal scheduling of MMG based on federated learning and reinforcement learning. Ref. [10] constructs an MMG system in cold and hot power areas, taking into account electric energy interaction. Although the above works have addressed the power interaction in a multi-microgrid system, the benefits of each MG are not considered enough. Regarding the aforementioned issue, some works have addressed the complexity of energy transactions between different entities in MMG systems. Ref. [11] considers an incompletely rational peer-to-peer MG energy transaction. Ref. [12] uses the particle swarm optimization (PSO) algorithm for peer-to-peer MMG economic dispatch. Ref. [13] proposes an MMG distributed power management in the shipping area based on an alternating direction method of multipliers (ADMM) algorithm. Ref. [14] leverages Monte Carlo simulations for the energy trading of MMG. Nevertheless, the high uncertainty, wide variability, and multi-energy coupling information of MMG systems present significant challenges in modeling the energy transactions between different entities. Currently, there are two main categories of methods for solving the MMG energy management scheduling model: model-driven approaches and data-driven approaches. Studies focusing on model-driven methods have also been conducted in this area. Ref. [15] proposes an improved genetic algorithm for MMG economic dispatch. Ref. [16] utilizes a PSO algorithm for the optimal scheduling of MG containing electric vehicles. Ref. [17] uses a distributed control method for the energy scheduling of MMG. Ref. [18] leverages the ADMM algorithm for the day-ahead scheduling of MMG based on the cooperative game model. Despite the progress made in this area, research on MMG energy management still faces several challenges due to the complexity of energy transactions between different entities and the uncertainty associated with renewable energy output. These challenges include: (1) The solution methods used heavily rely on the accuracy of the MMG model and lack robustness to the uncertainties associated with multiple available energy sources. Additionally, these methods may consume a significant amount of resources. (2) Moreover, the existing solution methods focus primarily on short-term benefits, neglecting the potential long-term benefits. Consequently, finding effective ways to address these challenges has become a key issue in MMG energy management.
To address the above challenges, we propose a data-driven approach that leverages a deep reinforcement learning (DRL) algorithm to coordinate the energy management of MMG. To be specific, deep neural networks avoid the dependence on precise mathematical equations and can automatically extract features from data to achieve precise model regression. In light of the high level of the uncertainty and limited data volume in MMG systems, reinforcement learning (RL) is suitable for real-time decision-making under complex and variable operating conditions [19]. In this way, Ref. [20] utilizes a fast online algorithm to solve the household load dispatching model, and the research results have achieved satisfactory results. Ref. [21] established an MG dispatch model considering renewable energy and used a hierarchical online algorithm to optimize the constructed objective function. However, the online optimization algorithm used in the above work has poor generalization performance compared to RL. Furthermore, the scheduling decision of the RL algorithm takes into account the potential impact of future long-term benefits, overcoming short-sightedness. However, existing RL methods are typically based on single-agent decision-making, which has limitations when dealing with complex scenarios. This is because single-agent RL relies on centralized scheduling and lacks autonomous learning capabilities, and may also face difficulties with the curse of dimensionality in complex multi-entity decision-making, resulting in convergence issues.
To address the aforementioned challenges, this paper proposes the use of multi-agent deep reinforcement learning (MADRL) for MMG optimal scheduling. Existing work has adopted MADRL to solve problems in power systems [22,23,24]. Ref. [25] proposes a layered hybrid MADRL to optimize a multi-service delivery business model that involves the coordination of multiple electric vehicles. Ref. [26] proposes a MADRL method for finding the optimal energy-saving strategy for hybrid electric vehicles. Ref. [27] utilizes the hybrid action space of MADRL to optimize off-grid building energy systems. Ref. [28] proposes the use of MADRL for the optimal scheduling of electric vehicle charging. However, the above works do not exploit the potential of MADRL to consider transactions between different entities. Furthermore, the generalization performance of the algorithm is also critical for practical applications. Ref. [29] uses the Nash-Q algorithm for multi-channel network system security control. However, this algorithm still faces the curse of dimensionality when dealing with complex scheduling scenarios, and its generalization ability is relatively poor, which limits its applicability. Moreover, this algorithm employs a discrete action space, which can lead to a reduction in calculation accuracy. In this regard, a soft actor-critic (SAC) algorithm, which combines value and policy iteration, has been successfully applied to power systems [30,31]. To tackle the energy management problem of an MMG system, which consists of multiple renewable energy microgrids belonging to different operating entities, we propose a multi-agent soft actor-critic (MASAC) algorithm that leverages automated machine learning (AutoML) as a skill to improve the generalization of MASAC and utilizes the experience replay buffer to reduce the temporal correlation between samples and improve training stability.
In summary, this paper proposes an MMG collaborative optimization scheduling model for MMG based on AutoML and MADRL to address the complex characteristics of the MMG system.
The main contributions of this study are summarized as follows:
  • To address the issue of the transaction and complementarity of electric energy among multi-microgrids, we constructed a collaborative optimization scheduling model for MMG based on a multi-agent centralized training distributed execution framework. This model effectively facilitates energy transactions between different entities and reduces the MMG system operating cost.
  • To enhance the generalization performance of the algorithm to cope with renewable energy uncertainties, we proposed an AutoML-based MASAC analysis method for MMG energy management. This approach eliminates the reliance on mathematical probability distributions for renewable energy outputs and increases the adaptability of the method to complex MMG scenarios.
  • Simulation tests have demonstrated that the proposed method can effectively manage the demand between different microgrids and promote the consumption of renewable energy, while achieving power complementarity. Moreover, the proposed method has better economy and computational efficiency than other RL algorithms.
  • The remaining sections of this paper are organized as follows: Section 2 mainly introduces the MMG energy management model, while Section 3 presents the solution method of the proposed model. In Section 4, we conduct a comprehensive case analysis to demonstrate the effectiveness of the proposed method. Finally, Section 5 summarizes the paper.

2. Multi-Microgrid Energy Management Model

The MMG system studied in this paper comprises multiple microgrids connected to the distribution network. Each individual MG can interact with other microgrids and trade energy with the distribution network through transmission lines. Before introducing the scheduling model of multi-microgrid in this section, we first discuss the individual MG model in detail.

2.1. Optimal Modeling of the Individual Microgrid

To clearly demonstrate the MG model, Figure 1 shows a schematic diagram of an individual MG’s structure, which is mainly composed of wind turbine (WT) units, photovoltaic (PV) units, electricity storage devices (ESD), micro-gas turbines (MGTS), load unit and an energy management center. Furthermore, the energy management center is responsible for the energy management of the MG.

2.1.1. Distributed Generation

The distributed generation in the studied MG includes WT and PV units. Following the principle of data-driven dispatch, this study employs real wind and photovoltaic power generation data for subsequent analysis, instead of modeling wind power and photovoltaic power generation output using explicit expressions [32].

2.1.2. Micro-Gas Turbines

The MGTS mainly burns natural gas to generate electricity, which can offer advantages such as high controllability and good power supply reliability. For the convenience of analysis, the cost function of the MGTS of MG i is set as follows:
C i t ( P M G T S , t , i ) = λ M G T S , M G   i P M G T S , t , i    t
P M G T S , i min P M G T S , t , i P M G T S , i max    t
where C i t ( P M G T S , t , i ) represents the operating cost of the MGTS of MG i at time t; P M G T S , t , i is the power generation of the MGTS of MG i at time t; λ M G T S , M G   i represent the power generation cost coefficients of MGTS of MG i; P M G T S , i min and P M G T S , i max are the minimum and maximum power generation of the MGTS of MG i, respectively.

2.1.3. Electricity Storage Devices

The ESD mainly achieves reasonable energy distribution by storing and releasing through the storage and release of electric energy, ultimately reducing the operating cost of the system [33]. At time t + 1, the relationship between the available capacity of the ESD and its charge and discharge power of MG i is expressed as:
S E S D , t + 1 , i = S E S D , t , i + ( η c h P c h , t , i P d c , t , i / η d c ) Δ t    t
where η c h and η d c represent the charging and discharging rate of the ESD, respectively; P c h , t , i and P d c , t , i are the charging and discharging power of the ESD of MG i in period t; S E S D , t , i and S E S D , t + 1 , i represent the capacity value of the ESD of MG i in period t and t + 1. Furthermore, we define the state of charge S O C E S D , t , i of the ESD of MG i at time t to detect the capacity of the ESD in real time as follows:
S O C E S D , t , i = S E S D , t , i / S E S D , max    t
where S E S D , max is the maximum capacity of the ESD. Furthermore, ESD improves the system economy through reasonable charge and discharge. For the convenience of analysis, the ESD operation and maintenance cost is set as follows:
C E S D , i ( t ) = ( | P c h , t , i | + | P d c , t , i | ) λ b    t
where C E S D , i ( t ) is the ESD operation and maintenance cost of MG i at time t; λ b is the operation and maintenance cost coefficient per unit power.

2.2. Multi-Microgrid Energy Management Model

To provide a clear representation of the MMG model, Figure 2 illustrates a schematic diagram of the MMG structure, which comprises multiple microgrids connected to the distribution network. The electricity trading process between MGs as well as between MGs and the distribution network is based on energy price information to ensure the economical operation of the MMG. Furthermore, the unified control center is responsible for managing and integrating MMG price information and system power requirements, which are then sent to individual MG. Additionally, the energy management center is responsible for the energy management of the MG based on the information provided by the unified control center.

2.2.1. Objective Function

The primary objective of MMG is to minimize the system’s operating cost. The operating cost of MMG consists primarily of the operating cost of micro-gas turbines, the transaction cost between MGs, the transaction cost between MGs and the distribution network, the operation and maintenance cost of ESD, the cost of active power loss, and the penalty cost of power imbalance between energy supply and consumption. Therefore, the objective function of MMG is:
min   C o s t = t = 1 T i = 1 n M G ( C i t ( P M G T S , t , i ) + C M G , i ( t ) + C G r i d , i ( t ) + C E S D , i ( t ) + λ l o s s P l o s s , t , i + ( P g a p , t , i ) 2 )
where n M G is the number of MGs; T is the number of time periods in a day; C i t ( P M G T S , t , i ) is the operating cost of MGTS of MG i at time t; C M G , i ( t ) is the transaction cost between the MG i and other MGs at time t; C G r i d , i ( t ) is the cost of transactions between MG i and distribution network at time t; C E S D , i ( t ) is the ESD operation and maintenance cost of MG i at time t; λ l o s s P l o s s , t , i is the loss cost of MG i in the process of energy transmission and generation side unit generation at time t; λ l o s s is the unit loss cost coefficient; P l o s s , t , i is the total power loss value of MG i during the energy transmission process and generation side unit generation at time t; ( P g a p , t , i ) 2 is the penalty cost of MG i in case of imbalance between energy supply and consumption at time t; P g a p , t , i is the power difference of MG i between the energy supplied and the energy consumed at time t; characterizes the penalty factor for the imbalance between energy supply and consumption.

Transaction Cost between Microgrids

The transaction cost between MGs is mainly determined by the price of electricity transacting between MGs as well as the amount of electricity traded. To reasonably arrange the transaction energy between MGs, the transaction cost of MG i is expressed as:
C M G , i ( t ) = j = 1 , j i n M G δ M G . t P i j , t    t
where δ M G . t is the purchase and sale price of electricity between MGs during the period t, and it is stipulated that the purchase price is equal to the electricity sale price; P i j , t is the transaction power between MGs i and j during period t; when the value is greater than 0, it is an electricity purchase, and when it is less than 0, it is for electricity sales.

Transaction Cost of MGs and Distribution Network

To reasonably arrange the electricity traded between the MGs and the distribution network as well as reduce the pressure on the power supply of the grid, the following transaction cost of MG i is set:
C G r i d , i ( t ) = δ G r i d , t P i g , t    t
where δ G r i d , t is the purchase and sale price of electricity between the MGs and distribution network during period t, and the stipulated electricity purchase price is greater than the electricity sale price; furthermore, the purchase and sale price between the MGs is between the purchase price and the sale price of the MGs and the distribution network; P i g , t is the electricity traded between the MG i and the distribution network during period t. If the value is greater than 0, it represents electricity purchase; on the contrary, it represents electricity sales.
To ensure the interests of the distribution network and to encourage energy transactions among MGs, we set the purchase and sale price of transactions between MGs to be lower than the purchase price between MGs and the distribution network; furthermore, the purchase and sale price between the MGs is between the purchase price and the sale price of the MGs and the distribution network. In case the MG experiences a shortage of power, it gives priority to purchasing power from other MGs. If the demand is still not met, the MG purchases power from the grid. Similarly, when the MG has surplus power and other MGs face a power shortage, it prioritizes meeting the load demand in the MMG system.

Microgrid Power Loss

The power loss taken into account in this study refers to the active power loss that occurs during the power generation of the generator set and energy transmission process. The generator-side unit comprises MGTS, WT, and PV. The formula used to calculate the specific power loss is as follows:
P l o s s , t , i = ψ M G T S , t , i P M G T S , t , i + ψ P V , t , i P P V , t , i + ψ W T , t , i P W T , t , i    t
ψ M G T S , t , i = P l o s s , t , i P M G T S , t , i    t
ψ P V , t , i = P l o s s , t , i P P V , t , i    t
ψ W T , t , i = P l o s s , t , i P W T , t , i    t
where ψ M G T S , t , i , ψ P V , t , i , ψ W T , t , i represent the power loss coefficients of micro-gas turbines, photovoltaics, and wind turbines, respectively; P W T , t , i is the power generated by the WT of MG i at time t; P P V , t , i is the power generated by PV of MG i at time t.

Power Imbalance between Energy Supply and Consumption

To facilitate the integration of renewable energy sources and achieve a balance between energy supply and demand, the unbalanced power of MG i is set to:
P g a p , t , i = P s u p , t , i P c o n , t , i    t
P s u p , t , i = P M G T S , t , i + P W T , t , i + P P V , t , i + P d c , t , i + P i j , t + P i g , t    t
P c o n , t , i = P l o a d , t , i + P c h , t , i + P l o s s , t , i    t
where P s u p , t , i is the energy provided by MG i at time t; P c o n , t , i is the energy consumed by MG i at time t; P l o a d , t , i is the load power of MG i at time t.

2.2.2. Constraints

Electrical Balance Constraint

To reasonably adjust the output of the power generation side and maintain the balance of energy supply and demand in the system, we set the following energy balance constraint of MG i:
P M G T S , t , i + P W T , t , i + P P V , t , i + P d c , t , i + P i j , t + P i g , t = P l o a d , t , i + P c h , t , i + P l o s s , t , i    t

Constraints of Electricity Storage Devices

To ensure that the charging and discharging power of ESD is within the allowable range, the limiting conditions of MG i are as follows [34]:
0 P c h , t , i P c h , max    t
0 P d c , t , i P d c , max    t
where P c h , max and P d c , max represent the maximum charging and discharging power of ESD.
To ensure that the ESD capacity is within the allowable range, the capacity of MG i must meet the following limits:
S E S D , min S E S D , t , i S E S D , max    t  
where S E S D , min is the minimum capacity of ESD.
Start and end limits: to ensure that the initial conditions remain consistent for each scheduling cycle, the ESD should adhere to the following start and end limits:
S 0 = S T , e n d = S E S D , min  
where S 0 = 0 and S T , e n d are the capacity of ESD at the beginning and end of the scheduling period T (in this work, T is taken as 24 h).

Constraints on Power Trading between MGs and Distribution Network

To avoid the excessive purchase of electricity from the distribution network, which may lead to higher electricity costs, the electricity traded between the MG i and the distribution network at time t is set as:
P i g , max P i g , t 0   P i g , t 0    t
0 P i g , t P i g , max    P i g , t 0    t
where P i g , max is the maximum power when the MG i trades with the grid.

Constraints on Electricity Traded between Microgrids

To prevent excessive power trading between MGs as well as avoid causing a supply–demand imbalance in MGs, we set the following power trading constraints:
P i j , t = P j i , t    t
0 P i j , t P i j , max    t
where P j i , t is the electricity traded between MG j and MG i at time t; P i j , max is the maximum power traded between MG i and MG j.

3. Model Solving

In this section, we first describe in detail the automated machine learning used to improve the generalization of MASAC, followed by a detailed introduction to the MASAC methodology proposed in this study.

3.1. Automated Machine Learning

Typically, the process of selecting neural network structures and hyperparameters for machine learning models involves a trial-and-error approach, which can be both tedious and challenging. To overcome this issue, we propose the use of complex control structures to operate machine learning models that can automatically learn appropriate parameters and configurations without the need for human intervention [35,36,37].
Optimizing hyperparameters for DRL algorithms is widely acknowledged as a complex task. In this study, we tackle this challenge by utilizing the currently popular AutoML technique to automatically find the best combination of hyperparameters for DRL. Figure 3 illustrates the structure of our approach. We use the metis tuner algorithm [38,39] to optimize hyperparameters. By leveraging metis to predict the next trial instead of guessing randomly, the AutoML finds the best hyperparameters for DRL. Moreover, we utilize AutoML to optimize the hyperparameters of MASAC with discount factor γ , actor network learning rate a _ l , critic network learning rate c _ l , mini-batch N, and adjustment coefficient .
Moreover, metis uses latin hypercube sampling (LHS) in stratified sampling [40], which divides the range of U parameters into D intervals and picks data points from the interval one at a time. Therefore, the number of combinations C of bootstrapping trials is
C = ( d = 0 D 1 D d ) U 1
After obtaining the above number of combinations, metis iteratively trains the Gaussian process model to enhance the robustness of tuning.

3.2. MASAC Methodology

Generally, the MADRL task can be described as a Markov decision process game (MDP game) [41]. Specifically, the MDP game consists of five key elements { [ S i ] n M G , [ A i ] n M G , [ ρ i ] n M G , [ R i ] n M G , [ γ i ] n M G } , where S i is the state set of agent i, [ S i ] n M G is the state set of all agents; A i represents the action set of agent i, and [ A i ] n M G is the action set of all agents; ρ i is the state transition matrix of agent i, [ ρ i ] n M G is the set of state transition matrices of all agents; R i is the return reward of agent i from state S i , t 1 to state S i , t , [ R i ] n M G is the reward set of all agents; γ i is the discount factor of agent i, which will affect the convergence of the algorithm; [ γ i ] n M G is the discount factor set of all agents. During the training process, the agent optimizes its own strategy, and the accumulated reward value gradually increases and tends to stabilize.
In this section, we apply the MDP game to the MMG scheduling model in this research. The key elements for each MG i are as follows.
(1) Agent: in each MG, the energy management center is set as an agent for the DRL algorithm.
(2) Environment: the environment is composed of PV, WT, ESD, loads, distribution network, and micro-gas turbines.
(3) State: the state is used to describe the environmental feedback of the action taken by the agent in the current environment. Specifically, the state includes the load power P l o a d , t , i of the MG i in period t, the state of charge S O C E S D , t , i of the electricity storage device of the MG i in period t, the power generation P W T , t , i of the WT of MG i in period t, the power generation P P V , t , i of the PV of MG i in period t, the transaction price δ M G , t between MGs in period t, as well as the transaction price information δ G r i d , t between MGs and the distribution network in period t. Therefore, the state set of agent i is:
S i = { P l o a d , t , i , S O C E S D , t , i , P W T , t , i , P P V , t , i , δ M G , t , δ G r i d , t }
(4) Action: action is mainly composed of the output P M G T S , t , i of the MGTS of the MG i at time t, the transaction strategy P i j , t between MG i and MG j at time t, and the transaction strategy P i g , t between MG i and the distribution network at time t. Therefore, the action set of agent i is:
A i = { P M G T S , t , i , P i j , t , P i g , t }
(5) Reward: the cost of each MG i includes the operation cost of micro-gas turbines, transaction cost with other MGs, transaction cost with the distribution network, ESD operation and maintenance cost, active power loss cost, and unbalanced power penalty cost. Furthermore, the goal of MG is to minimize the operating cost; therefore, the reward of agent i at time t is defined as follows:
R i ( t ) = ( C i t ( P M G T S , t , i ) + C M G , i ( t ) + C G r i d , i ( t ) + C E S D , i ( t ) + λ l o s s P l o s s , t , i + ( P g a p , t , i ) 2 )    t
In this regard, we define the total reward as the sum of the reward values of all agents. Moreover, in complex multi-agent interactive scenarios, the policy gradient method of the general single-agent reinforcement learning algorithm tends to increase the variance with the number of agents. Furthermore, the most single-agent reinforcement learning is a centralized learning method, which is not scalable. However, multi-agent reinforcement learning algorithms demonstrate superiority in multi-agent interaction scenarios. They acquire additional information during training to enhance stability, and the specific execution of the strategy depends only on the observation of the agent itself, without relying on the additional information. In this study, we adopt the MASAC algorithm, which follows the basic idea of centralized training and distributed execution (CTDE). Specifically, during training, the algorithm incorporates a global critic to guide actor training, while during testing, only actors with local observation environments are used to take action [42]. The advantage of this method is that it improves the efficiency of learning during training and improves the stability of training in a multi-agent environment. In this way, the framework of CTDE is shown in Figure 4.
Furthermore, we utilize the CTDE framework to extend the SAC algorithm to the dispatching scenario of multi-agent microgrids, which we call MASAC. This approach allows for the training of multiple agents in a high-dimensional continuous action space. The goal of MASAC is to maximize exploration by increasing entropy, thereby avoiding falling into a local optimal solution, and finding the strategy of global maximization. In MASAC, the actor of agent i updates the parameters of the policy network according to the gradient descent theory. The objective function of the specific policy network is as follows:
J ( φ i ) π i = E x [ κ log ( π φ i ( a ^ i | s i ) ) Q ξ , i ( x , a ^ ) ]
a ^ = { a ^ 1 , a ^ 2 , a ^ n M G }
x = { s 1 , s 2 , s n M G }
a = { a 1 , a 2 , a n M G }
r = { r 1 , r 2 , r n M G }
x = { s 1 , s 2 , s n M G }
where π φ i is the parameter of the actor network π of each agent i is φ ; Q ξ , i is the parameter of the critic network of agent i is ξ ; κ is the temperature parameter, which is used to control the influence ratio of entropy and reward; is the experience replay buffer, which is mainly used to store the joint state x , action a , reward r , and next state x ; a ^ is the action input into the critic network. Furthermore, the critic network of agent i updates the parameter ξ by minimizing the Bellman error J ( ξ i ) Q :
J ( ξ i ) Q = E x , a , r , x [ 1 2 ( Q ξ , i ( x , a ) w ) 2 ]
w = r i + γ E [ Q ξ i ¯ ( x , a ) κ log ( π φ i ( a i | s i ) ) ]
where ξ i ¯ is the target critic network parameter of agent i; a is the next action of agent i. During training, the actor and current critic network are utilized, while the target critic network performs parameter transfer from the current network to stabilize the training effect. After updating each critic network parameter, the target critic network parameter is soft updated, as follows:
ξ ¯ i = ϕ ξ i + ( 1 ϕ ) ξ i ¯
where ϕ is the hyperparameter controlling the soft update. Moreover, one of the main features of MASAC is the regularization of policy entropy. By increasing the exploration of actions, it can speed up the speed of the train and improve the quality of learning, preventing the policy from prematurely converging to a bad local optimal solution.
Moreover, the effective utilization of sampled data is also a key issue in MASAC. Experience replay buffer is a commonly used technique to store old and new experiences to prevent temporal correlations among samples [43], thereby improving the efficiency and quality of learning during training.
Based on the above analysis, Algorithm 1 summarizes the final MASAC algorithm.
Algorithm 1: MASAC Algorithm Based on AutoML for Multi-Microgrid Optimal Scheduling
1: Initialize the neural network parameters φ and ξ of actor and critic.
2: Initialize the replay buffer ℜ with size S .
3: for trial = 1: M do
4:  Select a set of hyperparameters from the search space according to the Metis Tuner.
5:  for episode = 1: E do
6:    Select random action from the action space.
7:    Select the initial state from the state space.
8:    for t = 1: H do
9:     Each agent i selects action ai from the action space.
10:      Interact joint actions a = {a1,a2,… anMG) with the environment to get corresponding states x′ and rewards r.
11:      Store transition (x,a,r,x′) in experience replay buffer ℜ.
12:      for agent = 1: nMG do
13:       Sample a mini-batch of N experience (xN,aN,rN, x N ) from the experience replay buffer ℜ.
14:       Updating the critic network by minimizing the loss function.
15:       Updating the actor network via gradient descent.
16:      end for
17:      Updating critic target network parameters using soft update.
18:     end for
19:    end for
20:  Collect the reward and upload it to the Metis Tuner.
21: end for
22: Select the best hyperparameters and policies.

3.3. Solving Process

The solution process of the MMG scheduling model is as follows:
Step 1: Construct a scheduling model according to the Formulas (6)–(24).
Step 2: Input the MMG parameters.
Step 3: Set and update the episode of MASAC training.
Step 4: According to the state set and action set, calculate the reward function according to the Formula (28).
Step 5: Determine whether a solution exists. If it exists and meets the stopping criteria, the process terminates; otherwise, return to Step 2.
Step 6: Obtain the optimal scheduling strategy for MMG.

4. Case Study

To verify the effectiveness of the proposed scheduling model and method, the following simulation experiments are carried out. Moreover, the MASAC algorithm proposed in this study has been implemented in Python 3.8 using Pytorch 1.10. All simulation tests are carried out on a PC platform equipped with Intel Core i5-6300HQ CPU (2.3 GHz) and 8 GB RAM.

4.1. Settings in Test Case

In this study, we set up a test case of an MMG system test case consisting of two microgrids. The key components of the MMG system include micro-gas turbines, wind turbines, photovoltaics, electricity storage devices, loads, and energy management centers. Furthermore, the data records for WT power generation are provided by Fortum Oyj from a wind farm in Finland, while the data related to PVs are obtained from [44]. Moreover, the time range of the simulation test is set to T = 24 h, and the time interval is t = 1 h. Figure 5 and Figure 6 are the data curves of WT and PV power generation and load power of MG 1 and MG 2, respectively. It can be seen from the data that the load value of MG 1 is higher than the output of renewable energy in multiple periods, while the load value of MG 2 is lower than the output of renewable energy in multiple periods. Figure 7 depicts the price information for transactions between the MGs and the grid as well as between MGs. Moreover, the maximum trading power between the MGs and the grid is 500 kW, and the maximum trading power between MGs is 200 kW. Table 1 describes the main parameter settings of the MG [45]. Furthermore, the main implementation details of MASAC are as follows: Specifically, the basic structure of the neural network of each MG is consistent, and the Adam optimizer is used. The γ of MASAC is 0.916, the learning rates of actor and critic are a _ l = 0.0004 and c _ l = 0.0006, respectively, the size of the experience replay buffer S is 10,000, is 0.159, and the sampling mini-batch N is 512.

4.2. Results and Analysis

4.2.1. Analyze Optimization Results Using AutoML

To evaluate the effectiveness of AutoML, the following simulation tests have been carried out. AutoML assesses the intermediate results generated by the current hyperparameter selection and offers reasonable suggestions for the next hyperparameter trajectory. Finally, the hyperparameter selection results for each trajectory are displayed on the WebUI, as shown in Figure 8.
Figure 8 shows the results of AutoML’s optimization of the hyperparameters required by MASAC. In this figure, each curve represents a set of hyperparameters for a trial, each ordinate represents a range of hyperparameters, and the last ordinate is the reported total reward value for all agents using those hyperparameters. Furthermore, the darker the red, the more appropriate the hyperparameter set, and the green color indicates that the selected combination of hyperparameters cannot achieve satisfactory results.
Furthermore, to validate the rationality of AutoML’s multiple trial results, the following simulations were conducted. Figure 9 shows the results of the final report total reward value for all agents in multiple trials using AutoML, where each coordinate point represents a trial. It can be seen from the figure that except for a few trial results that deviate from the normal value of the trial, most of the trials can achieve satisfactory results. Based on the final reported total reward value for all agents, it is evident that the designed AutoML can achieve desirable optimization results.
Upon analyzing the results above, it is evident that AutoML is capable of selecting the optimal combination of hyperparameters for MASAC, leading to an improvement in the algorithm’s generalization ability and learning efficiency.
Furthermore, to verify the adaptability of AutoML to parameter adjustment when the input parameters are changed, we changed the values of the input parameters in Table 2 and performed simulation experiments. Note that Table 2 here is only a set of parameters for verifying the adaptive setting of the above-mentioned AutoML, which is different from the parameters of the MMG optimization model constructed in Table 1. Furthermore, we set up two experiments as follows:
Experiment 1: Optimizing the scheduling model under the input parameters in Table 1.
Experiment 2: Optimizing the scheduling model under the input parameters in Table 2.
Table 3 shows the best hyperparameter results optimized by AutoML for MASAC under different experiments. It can be seen from the table that under different input parameters, the optimal combination of hyperparameters optimized by the experiment is different. This shows that AutoML can automatically select the best hyperparameters for MASAC according to different inputs to formulate a reasonable scheduling strategy. Additionally, the AutoML adaptability was verified.

4.2.2. Electrical Balance Analysis of Each MG

To verify the electrical balance effect of each MG, we conducted a simulation analysis, and the results are shown in Figure 10 and Figure 11. The analysis demonstrates that each MG attains its own energy supply balance by means of energy transactions with other MGs and the distribution network.
It is evident from Figure 10 that the WT and PV outputs of MG 1 are lower than the power of the electric load during the peak period of power consumption, resulting in a power deficit for this MG. As a result, MG 1 receives additional power supply from other MGs and the distribution network to meet the demand. Similarly, Figure 11 shows that the WT and PV outputs of MG 2 are higher than the power of the electric load in most periods of time, indicating a power surplus MG. The excess electricity produced can be sold to other MGs or the distribution network, resulting in additional income. Specifically, during the off-peak period of electricity consumption, each MG charges the excess electricity to its own ESD, which is then discharged during the peak period of electricity consumption. Additionally, the excess electricity is sold to other MGs and the distribution network to generate further revenue. During peak hours of power consumption, MG 1 discharges its own ESD first. If the energy supplied by the MG itself is insufficient, it purchases electricity from other power surplus MGs at a lower transaction price than the price between MGs and the distribution network. If the energy demand is still not met, electricity is purchased from the grid to maintain the energy supply and demand balance. Through the energy complementarity between the MGs, the full utilization of energy is realized, and the power supply pressure of the distribution network is reduced. Additionally, the electrical balance of each MG is achieved.

4.2.3. Economic Analysis

In order to verify the effectiveness of the proposed multi-microgrid scheduling model, the following two modes are set and simulation experiments are carried out.
Model 1: Transactions between microgrids are not considered in a single MG, only transactions between microgrids and distribution network are considered.
Model 2: Consider transactions between microgrids as well as between microgrids and distribution network in a single MG.
Table 4 shows the operating cost of MMG in two different modes. Furthermore, the MMG in this study works in Mode 2. It can be seen from Table 4 that the cost of MMG under Mode 2 is reduced by 7.36% compared with that under Mode 1. This shows that the power interaction between microgrids can effectively reduce the operating cost of the MMG system and improve the economics of system operation.

4.2.4. Analysis of Transactions between MGs as Well as between MGs and the Distribution Network

To verify the effectiveness of the proposed scheduling scheme, simulation results were tested and presented in Figure 12, which describes the process of energy transactions between MGs as well as between MGs and the grid. The simulation results reveal that the power purchase price between MGs is much lower than that of MGs and the grid, as well as the price of electricity sold between MGs is higher than the price of electricity sold between MGs and the grid, hence the priority of energy transactions between MGs is higher than that of MGs and the grid. Specifically, when MG 2 has sufficient power, it gives priority to selling the excess power to the power shortage MG 1, and then finally sells the remaining energy to the grid. Similarly, when the power shortage MG 2 cannot meet the energy demand, it gives priority to purchasing power from MG 1 and then from the grid. The above analysis shows that the MMG system optimally utilizes energy and achieves a high economic utilization of energy through the energy complementation of each MG.

4.2.5. ESD Charging and Discharging Strategy Analysis

To verify the effectiveness of the ESD charging and discharging strategy, a simulation test was conducted and the results are shown in Figure 13. As can be observed from the figure, both MG 1 and MG 2’s ESD store enough power when the power is sufficient; when the power consumption peaks and the power is short, the ESD releases the stored energy. The power released by the ESD reduces the power purchased by the MG from the grid, further relieving the pressure on the grid power supply. Additionally, the charging and discharging strategy of ESD considers the energy shortage in future peak hours, effectively increasing the MMG system’s operating flexibility.

4.2.6. Performance Comparison with other RL Algorithms

To show the superiority of the proposed method, it has been tested in comparison with other RL methods. Figure 14 illustrates the convergence of different RL algorithms. During the initial learning phase, each RL method explores different directions randomly, which may not result in a more profitable policy and consequently leads to a lower total reward value for all agents. However, as the accumulated experience increases, the total reward value of all RL methods starts to increase continuously and eventually converges.
Furthermore, the proposed method exhibits significantly higher total reward value for all agents compared to other RL algorithms, indicating a lower operating cost. Here, the proximal policy optimization (PPO) and advantage actor-critic (A2C) algorithms are extended to the multi-agent space in this study, named MAPPO and MAA2C, respectively. Based on the results shown in Figure 14, it can be observed that the proposed method has reduced operating costs by 12.9% and 17.30% compared to MAPPO and MAA2C, respectively. The reasons for this are analyzed as follows: (1) the experience replay buffer and AutoML have improved the stability of training, and (2) MASAC has effectively enhanced the generalization performance through the framework of maximum entropy and CTDE. Hence, it can be concluded that our proposed method is more economical compared to other RL methods.
Additionally, to verify the convergence performance of the proposed method, we conducted simulation tests and present the results below. Table 5 compares the number of episodes and the calculation time required for different RL algorithms to reach convergence. Table 5 clearly indicates that the proposed method has the shortest convergence time and requires the least number of episodes to converge, as compared to other RL algorithms. Thanks to the strong generalization of the MASAC, it can quickly identify the optimal strategy and converge and stabilize faster. Thus, the proposed method outperforms other RL algorithms in terms of computational performance.

4.2.7. Analysis of Convergence and Computational Efficiency across Multiple Runs

In order to verify the convergence performance and computational efficiency of the proposed method for multiple runs with the same input parameters, the following simulation experiments are performed.
Table 6 presents the total reward value for all agents and computation time in multiple runs with the same input parameters. Here, each run comprises 1000 episodes. It can be seen from the table that the total reward value for all agents has not changed across multiple runs, and the difference between the longest calculation time and the shortest calculation time during the training process is 12.01 s, which falls within an acceptable range. Additionally, the average computation time over 10 runs is 446.17 s. This shows that the proposed method can converge stably across multiple runs.

5. Conclusions

To investigate the complementarity and trading of electric energy between MGs under different operating entities in MMG, this study proposes a multi-agent deep reinforcement learning scheduling method. The proposed method employs a multi-agent centralized training distributed execution framework to address uncertainty in the environment and determine the optimal trading strategy. Based on the simulation results, we draw the following conclusions.
(1) The established MMG scheduling model, based on a multi-agent centralized training distributed execution framework, allows for each MG to adjust the power interaction between MGs based on energy transaction prices and energy demand. This helps to reduce the cost of energy utilization and dependence on grid energy supply, while effectively facilitating energy transactions between different entities and improving the economics of MMG system operation.
(2) The developed MASAC algorithm, based on automated machine learning, has been shown to be capable of addressing the collaborative optimal scheduling of MMG and achieving satisfactory convergence by learning from historical experience. This approach is better suited to complex scheduling scenarios and real-time online scheduling decisions.
(3) The test results prove that the proposed method is more economical and computationally efficient than other RL algorithms.
In future work, various flexible loads will be considered to increase the flexibility on the demand side. It is interesting to extend this work to include the dispatch of multiple energy forms, such as heat and power, in an integrated energy system [46,47]. Another topic worthy of research is the resilient scheduling of information attacks [48].

Author Contributions

J.G.: data curation, methodology, formal analysis, software, writing—original draft. Y.L.: methodology, data curation, investigation, supervision, writing—review and editing. B.W.: methodology, resources, visualization, writing—review and editing. H.W.: visualization, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the data being confidential.

Acknowledgments

This work received no external funding support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Parlikar, A.; Schott, M.; Godse, K.; Kucevic, D.; Jossen, A.; Hesse, H. High-power electric vehicle charging: Low-carbon grid integration pathways with stationary lithium-ion battery systems and renewable generation. Appl. Energy 2023, 333, 120541. [Google Scholar] [CrossRef]
  2. Li, Y.; Han, M.; Shahidehpour, M.; Li, J.Z.; Long, C. Data-driven distributionally robust scheduling of community integrated energy systems with uncertain renewable generations considering integrated demand response. Appl. Energy 2023, 335, 120749. [Google Scholar] [CrossRef]
  3. Kim, H.J.; Kim, M.K. A novel deep learning-based forecasting model optimized by heuristic algorithm for energy management of microgrid. Appl. Energy 2023, 332, 120525. [Google Scholar] [CrossRef]
  4. Feng, Z.N.; Wei, F.R.; Wu, C.T.; Sui, Q.; Lin, X.N.; Li, Z.T. Novel source-storage coordination strategy adaptive to impulsive generation characteristic suitable for isolated island microgrid scheduling. IEEE Trans. Smart Grid 2023, 2023, 3244852. [Google Scholar] [CrossRef]
  5. Li, Y.; Yang, Z.; Li, G.Q.; Zhao, D.B.; Tian, W. Optimal scheduling of an isolated microgrid with battery storage considering load and renewable generation uncertainties. IEEE Trans. Ind. Electr. 2018, 66, 1565–1575. [Google Scholar] [CrossRef] [Green Version]
  6. An, D.; Yang, Q.Y.; Li, D.H.; Wu, Z.Z. Distributed Online Incentive Scheme for Energy Trading in Multi-Microgrid Systems. IEEE Trans. Autom. Sci. Eng. 2023, 2023, 3236408. [Google Scholar] [CrossRef]
  7. Hakimi, S.M.; Hasankhani, A.; Shafie-khah, M.; Catalão, J.P.S. Stochastic planning of a multi-microgrid considering integration of renewable energy resources and real-time electricity market. Appl. Energy 2021, 298, 117215. [Google Scholar] [CrossRef]
  8. Zhao, Z.L.; Guo, J.T.; Luo, X.; Lai, C.S.; Yang, P.; Lai, L.L.; Li, P.; Guerrero, J.M.; Shahidehpour, M. Distributed robust model predictive control-based energy management strategy for islanded multi-microgrids considering uncertainty. IEEE Trans. Smart Grid 2022, 13, 2107–2120. [Google Scholar] [CrossRef]
  9. Li, Y.Z.; He, S.Y.; Li, Y.; Shi, Y.; Zeng, Z.G. Federated multiagent deep reinforcement learning approach via physics-informed reward for multimicrogrid energy management. IEEE Trans. Neur. Netw. Learn. Syst. 2023, 2023, 3232630. [Google Scholar] [CrossRef]
  10. Lin, S.F.; Liu, C.T.; Li, D.D.; Fu, Y. Bi-level multiple scenarios collaborative optimization configuration of CCHP regional multi-microgrid system considering power interaction among microgrids. Proc. CSEE 2020, 40, 1409–1421. [Google Scholar] [CrossRef]
  11. Xia, Y.X.; Xu, Q.S.; Huang, Y.; Liu, Y.H.; Li, F.X. Preserving privacy in nested peer-to-peer energy trading in networked microgrids considering incomplete rationality. IEEE Trans. Smart Grid 2022, 14, 606–622. [Google Scholar] [CrossRef]
  12. Ali, L.; Muyeen, S.M.; Bizhani, H.; Simoes, M.G. Economic planning and comparative analysis of market-driven multi-microgrid system for peer-to-peer energy trading. IEEE Trans. Ind. Appl. 2022, 58, 4025–4036. [Google Scholar] [CrossRef]
  13. Xie, P.L.; Tan, S.; Bazmohammadi, N.; Guerrero, J.M.; Vasquez, J.C.; Alcala, J.M.; Carreño, J.E.M. A distributed real-time power management scheme for shipboard zonal multi-microgrid system. Appl. Energy 2022, 317, 119072. [Google Scholar] [CrossRef]
  14. Daneshvar, M.; Mohammadi-Ivatloo, B.; Abapour, M.; Asadi, S. Energy exchange control in multiple microgrids with transactive energy management. J. Mod. Power Syst. Clean Energy 2020, 8, 719–726. [Google Scholar] [CrossRef]
  15. Jiang, H.Y.; Ning, S.Y.; Ge, Q.B.; Yun, W.; Xu, J.Q.; Bin, Y. Optimal economic dispatching of multi-microgrids by an improved genetic algorithm. IET Cyber-Syst. Robot. 2021, 3, 68–76. [Google Scholar] [CrossRef]
  16. Zhang, X.Z.; Wang, Z.Y.; Lu, Z.Y. Multi-objective load dispatch for microgrid with electric vehicles using modified gravitational search and particle swarm optimization algorithm. Appl. Energy 2022, 306, 118018. [Google Scholar] [CrossRef]
  17. Nawaz, A.; Wu, J.; Ye, J.; Dong, Y.D.; Long, C.N. Distributed MPC-based energy scheduling for islanded multi-microgrid considering battery degradation and cyclic life deterioration. Appl. Energy 2023, 329, 120168. [Google Scholar] [CrossRef]
  18. Chen, W.D.; Wang, J.N.; Yu, G.Y.; Chen, J.J.; Hu, Y.M. Research on day-ahead transactions between multi-microgrid based on cooperative game model. Appl. Energy 2022, 316, 119106. [Google Scholar] [CrossRef]
  19. Li, Y.; Bu, F.J.; Li, Y.Z.; Long, C. Optimal scheduling of island integrated energy systems considering multi-uncertainties and hydrothermal simultaneous transmission: A deep reinforcement learning approach. Appl. Energy 2023, 333, 120540. [Google Scholar] [CrossRef]
  20. Alahyari, A.; Jooshaki, M. Fast energy management approach for the aggregated residential load and storage under uncertainty. J. Energy Storage 2023, 62, 106848. [Google Scholar] [CrossRef]
  21. Zou, H.L.; Wang, Y.; Mao, S.W.; Zhang, F.H.; Chen, X. Distributed online energy management in interconnected microgrids. IEEE Intern. Things J. 2019, 7, 2738–2750. [Google Scholar] [CrossRef]
  22. Fan, Z.; Zhang, W.; Liu, W.X. Multi-agent deep reinforcement learning based distributed optimal generation control of DC microgrids. IEEE Trans. Smart Grid 2023, 2023, 3237200. [Google Scholar] [CrossRef]
  23. Wang, Y.; Qiu, D.W.; Teng, F.; Strbac, G. Towards microgrid resilience enhancement via mobile power sources and repair crews: A multi-agent reinforcement learning approach. IEEE Trans. Power Syst. 2023, 2023, 3240479. [Google Scholar] [CrossRef]
  24. Hu, C.F.; Wen, G.H.; Wang, S.; Fu, J.J.; Yu, W.W. Distributed Multiagent Reinforcement Learning with Action Networks for Dynamic Economic Dispatch. IEEE Trans. Neur. Netw. Learn. Syst. 2023, 2023, 3234049. [Google Scholar] [CrossRef]
  25. Qiu, D.W.; Wang, Y.; Sun, M.Y.; Strbac, G. Multi-service provision for electric vehicles in power-transportation networks towards a low-carbon transition: A hierarchical and hybrid multi-agent reinforcement learning approach. Appl. Energy 2022, 313, 118790. [Google Scholar] [CrossRef]
  26. Wang, Y.; Wu, Y.K.; Tang, Y.J.; Li, Q.; He, H.W. Cooperative energy management and eco-driving of plug-in hybrid electric vehicle via multi-agent reinforcement learning. Appl. Energy 2023, 332, 120563. [Google Scholar] [CrossRef]
  27. Gao, Y.; Matsunami, Y.; Miyata, S.; Akashi, Y. Multi-agent reinforcement learning dealing with hybrid action spaces: A case study for off-grid oriented renewable building energy system. Appl. Energy 2022, 326, 120021. [Google Scholar] [CrossRef]
  28. Park, K.; Moon, I. Multi-agent deep reinforcement learning approach for EV charging scheduling in a smart grid. Appl. Energy 2022, 328, 120111. [Google Scholar] [CrossRef]
  29. Yu, Y.; Liu, G.P.; Hu, W.S. Learning-based secure control for multichannel networked systems under smart attacks. IEEE Trans. Ind. Electr. 2022, 70, 7183–7193. [Google Scholar] [CrossRef]
  30. Xia, Y.; Xu, Y.; Wang, Y.; Mondal, S.; Dasgupta, S.; Gupta, A.K.; Gupta, G.M. A safe policy learning-based method for decentralized and economic frequency control in isolated net-worked-microgrid systems. IEEE Trans. Sustain. Energy 2022, 13, 1982–1993. [Google Scholar] [CrossRef]
  31. Soleimanzade, M.A.; Kumar, A.; Sadrzadeh, M. Novel data-driven energy management of a hybrid photovoltaic-reverse osmosis desalination system using deep reinforcement learning. Appl. Energy 2022, 317, 119184. [Google Scholar] [CrossRef]
  32. Li, Y.; Wang, R.N.; Li, Y.Z.; Zhang, M.; Long, C. Wind power forecasting considering data privacy protection: A federated deep reinforcement learning approach. Appl. Energy 2023, 329, 120291. [Google Scholar] [CrossRef]
  33. Li, Y.; Wang, B.; Yang, Z.; Li, J.Z.; Li, G.Q. Optimal scheduling of integrated demand response-enabled community-integrated energy systems in uncertain environments. IEEE Trans. Ind. Appl. 2021, 58, 2640–2651. [Google Scholar] [CrossRef]
  34. Li, Y.; Feng, B.; Wang, B.; Sun, S.C. Joint planning of distributed generations and energy storage in active distribution networks: A Bi-Level programming approach. Energy 2022, 245, 123226. [Google Scholar] [CrossRef]
  35. Hutter, F.; Kotthoff, L.; Vanschoren, J. Automated Machine Learning: Methods, Systems, Challenges; Springer: Cham, Switzerland, 2019. [Google Scholar]
  36. Awad, M.; Khanna, R. Machine Learning and Knowledge Discovery. In Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers; Awad, M., Khanna, R., Eds.; Apress: Berkeley, CA, USA, 2015; pp. 19–38. [Google Scholar]
  37. He, X.; Zhao, K.Y.; Chu, X.W. AutoML: A survey of the state-of-the-art. Knowl. Based Syst. 2021, 212, 106622. [Google Scholar] [CrossRef]
  38. Li, Z.L.; Liang, C.J.M.; He, W.J.; Zhu, L.J.; Dai, W.J.; Jiang, J.; Sun, G.Z. Metis: Robustly tuning tail latencies of cloud systems. In Proceedings of the 2018 USENIX Annual Technical Conference (USENIX ATC ’18), Boston, MA, USA, 11–13 July 2018; pp. 981–992. [Google Scholar]
  39. Li, Y.; Wang, R.N.; Yang, Z. Optimal scheduling of isolated microgrids using automated reinforcement learning-based multi-period forecasting. IEEE Trans. Sustain. Energy 2021, 13, 159–169. [Google Scholar] [CrossRef]
  40. McKay, M.D.; Beckman, R.J.; Conover, W.J. A comparison of three methods for selecting values of input variables in the analysis of output from a computer code. Technometrics 2000, 42, 55–61. [Google Scholar] [CrossRef]
  41. Buşoniu, L.; Babuška, R.; De Schutter, B. Multi-Agent Reinforcement Learning: An Overview. In Innovations in Multi-Agent Systems and Applications-1; Srinivasan, D., Jain, L.C., Eds.; Studies in Computational Intelligence; Springer: Berlin, Heidelberg, 2010; Volume 310, pp. 183–221. [Google Scholar]
  42. Lowe, R.; Wu, Y.I.; Tamar, A.; Harb, J.; Pieter Abbeel, O.; Mordatch, I. Multi-agent actor-critic for mixed cooperative-competitive environments. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 1–12. [Google Scholar]
  43. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  44. Logenthiran, T.; Srinivasan, D.; Khambadkone, A.M.; Aung, H.N. Multiagent system for real-time operation of a microgrid in real-time digital simulator. IEEE Trans. Smart Grid 2012, 3, 925–933. [Google Scholar] [CrossRef]
  45. Li, Y.; Wang, B.; Yang, Z.; Li, J.Z.; Chen, C. Hierarchical stochastic scheduling of multi-community integrated energy systems in uncertain environments via Stackelberg game. Appl. Energy 2022, 308, 118392. [Google Scholar] [CrossRef]
  46. Sun, J.Z.; Deng, J.H.; Li, Y. Indicator & crowding distance-based evolutionary algorithm for combined heat and power economic emission dispatch. Appl. Soft Comput. 2020, 90, 106158. [Google Scholar] [CrossRef] [Green Version]
  47. Zhang, Z.; Chen, Y.B.; Ma, J.; Zhao, D.W.; Qian, M.H.; Li, D.; Wang, D.; Zhao, L.H.; Zhou, M. Stochastic optimal dispatch of combined heat and power integrated AA-CAES power station considering thermal inertia of DHN. Int. J. Electr. Power Energy Syst. 2022, 141, 108151. [Google Scholar] [CrossRef]
  48. Li, Y.; Wei, X.H.; Li, Y.Z.; Dong, Z.Y.; Shahidehpour, M. Detection of false data injection attacks in smart grid: A secure federated deep learning approach. IEEE Trans. Smart Grid 2022, 13, 4862–4872. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of an individual MG structure.
Figure 1. Schematic diagram of an individual MG structure.
Energies 16 03248 g001
Figure 2. Schematic diagram of the MMG structure.
Figure 2. Schematic diagram of the MMG structure.
Energies 16 03248 g002
Figure 3. AutoML-based MASAC hyperparameter optimization.
Figure 3. AutoML-based MASAC hyperparameter optimization.
Energies 16 03248 g003
Figure 4. The CTDE structure diagram.
Figure 4. The CTDE structure diagram.
Energies 16 03248 g004
Figure 5. WT, PV output, and load power curve of MG 1.
Figure 5. WT, PV output, and load power curve of MG 1.
Energies 16 03248 g005
Figure 6. WT, PV output, and load power curve of MG 2.
Figure 6. WT, PV output, and load power curve of MG 2.
Energies 16 03248 g006
Figure 7. Transaction prices between MGs and grid as well as between MGs.
Figure 7. Transaction prices between MGs and grid as well as between MGs.
Energies 16 03248 g007
Figure 8. Results of hyperparameter optimization using AutoML.
Figure 8. Results of hyperparameter optimization using AutoML.
Energies 16 03248 g008
Figure 9. Total reward distribution under multiple trials.
Figure 9. Total reward distribution under multiple trials.
Energies 16 03248 g009
Figure 10. Electrical balance of MG 1.
Figure 10. Electrical balance of MG 1.
Energies 16 03248 g010
Figure 11. Electrical balance of MG 2.
Figure 11. Electrical balance of MG 2.
Energies 16 03248 g011
Figure 12. Figure 12. Electric power trading.
Figure 12. Figure 12. Electric power trading.
Energies 16 03248 g012
Figure 13. ESD charging and discharging strategy.
Figure 13. ESD charging and discharging strategy.
Energies 16 03248 g013
Figure 14. Comparison of different RL algorithm’s total reward.
Figure 14. Comparison of different RL algorithm’s total reward.
Energies 16 03248 g014
Table 1. Main parameter settings of each MG.
Table 1. Main parameter settings of each MG.
ParameterValueParameterValue
P c h , max (kW)100 λ M G T S , M G   1 ($/kWh)1.3
P d c , max (kW)100 λ M G T S , M G   2 ($/kWh)1.5
S E S D , max (kWh)200 λ l o s s ($/kWh)1.35
η c h , η d c 0.9 0.5
P M G T S , i min (kW)5 n M G 2
P M G T S , i max (kW)30 ψ M G T S , t , i , ψ P V , t , i , ψ W T , t , i 0.02
λ b ($/kWh)0.5
Table 2. A set of parameters to verify the adaptability of AutoML.
Table 2. A set of parameters to verify the adaptability of AutoML.
ParameterValueParameterValue
P c h , max (kW)100 λ M G T S , M G   1 ($/kWh)0.1
P d c , max (kW)100 λ M G T S , M G   2 ($/kWh)0.2
S E S D , max (kWh)200 λ l o s s ($/kWh)0.15
η c h , η d c 0.9 0.5
P M G T S , i min (kW)5 n M G 2
P M G T S , i max (kW)30 ψ M G T S , t , i , ψ P V , t , i , ψ W T , t , i 0.02
λ b ($/kWh)0.06
Table 3. Results of AutoML optimization hyperparameters under different experiments.
Table 3. Results of AutoML optimization hyperparameters under different experiments.
γ a _ l c _ l N
Experiment 10.9160.00040.00065120.159
Experiment 20.8770.00020.00071280.269
Table 4. Operating cost of MMG under different modes.
Table 4. Operating cost of MMG under different modes.
Model 1 (USD)Model 2 (USD)
MMG63624.0058942.00
Table 5. Convergence comparison of different RL algorithms.
Table 5. Convergence comparison of different RL algorithms.
Solution MethodNumber of EpisodesConvergence Time (s)
Proposed method545236.43
MAPPO771267.67
MAA2C995339.24
Table 6. Total reward value for all agents and computation time in multiple runs.
Table 6. Total reward value for all agents and computation time in multiple runs.
Run NumberTotal Reward (10−4)Computation Time (s)
1−5.89442.00
2−5.89449.11
3−5.89451.78
4−5.89447.33
5−5.89444.08
6−5.89439.77
7−5.89447.56
8−5.89439.99
9−5.89449.42
10−5.89450.68
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, J.; Li, Y.; Wang, B.; Wu, H. Multi-Microgrid Collaborative Optimization Scheduling Using an Improved Multi-Agent Soft Actor-Critic Algorithm. Energies 2023, 16, 3248. https://0-doi-org.brum.beds.ac.uk/10.3390/en16073248

AMA Style

Gao J, Li Y, Wang B, Wu H. Multi-Microgrid Collaborative Optimization Scheduling Using an Improved Multi-Agent Soft Actor-Critic Algorithm. Energies. 2023; 16(7):3248. https://0-doi-org.brum.beds.ac.uk/10.3390/en16073248

Chicago/Turabian Style

Gao, Jiankai, Yang Li, Bin Wang, and Haibo Wu. 2023. "Multi-Microgrid Collaborative Optimization Scheduling Using an Improved Multi-Agent Soft Actor-Critic Algorithm" Energies 16, no. 7: 3248. https://0-doi-org.brum.beds.ac.uk/10.3390/en16073248

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop