Next Article in Journal
Diagnosing Complex Organisations with Diverse Cultures—Part 2: Application to ASEAN
Previous Article in Journal
Setting the Public Sentiment: Examining the Relationship between Social Media and News Sentiments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning-Based Multi-Objective Optimization for Generation Scheduling in Power Systems

1
Major in Industrial Data Science & Engineering, Department of Industrial and Data Engineering, Pukyong National University, Busan 48513, Republic of Korea
2
Department of Systems Management and Engineering, Pukyong National University, Busan 48513, Republic of Korea
*
Author to whom correspondence should be addressed.
Systems 2024, 12(3), 106; https://doi.org/10.3390/systems12030106
Submission received: 22 February 2024 / Revised: 11 March 2024 / Accepted: 18 March 2024 / Published: 19 March 2024

Abstract

:
Multi-objective power scheduling (MOPS) aims to address the simultaneous minimization of economic costs and different types of environmental emissions during electricity generation. Recognizing it as an NP-hard problem, this article proposes a novel multi-agent deep reinforcement learning (MADRL)-based optimization algorithm. Within a custom multi-agent simulation environment, representing power-generating units as collaborative types of reinforcement learning (RL) agents, the MOPS problem is decomposed into sequential Markov decision processes (MDPs). The MDPs are then utilized for training an MADRL model, which subsequently offers the optimal solution to the optimization problem. The practical viability of the proposed method is evaluated across several experimental test systems consisting of up to 100 units featuring bi-objective and tri-objective problems. The results demonstrate that the proposed MADRL algorithm has better performance compared to established methods, such as teaching learning-based optimization (TLBO), real coded grey wolf optimization (RCGWO), evolutionary algorithm based on decomposition (EAD), non-dominated sorting algorithm II (NSGA-II), and non-dominated sorting algorithm III (NSGA-III).

1. Introduction

Due to the instantaneous nature of electrical power and its inability to be stored in massive quantities, its demand exhibits temporal fluctuations. This dynamic nature of electrical consumption prompts the need for reliable power scheduling [1]. Central to power scheduling is the fundamental requirement that the power supply match the demand at each period of a particular planning horizon. Power system operators utilize a set of generating units to achieve this equilibrium, each with its own technological and operational constraints. The cost of power generation also varies among generating units, and it is not directly proportional to the power output [2]. Economic cost dispatch (ECD) forms the foundational aspect of power scheduling, representing a single-objective optimization problem. Subject to various unit-specific and system-level constraints, ECD determines the optimal loads of generating units that minimize the operating costs. This process requires a delicate balance among fluctuating demands, operational constraints, and economic considerations. There are also growing concerns about the environmental impact of power generation because approximately two-thirds of electric energy is currently derived from fossil fuels [3]. Fossil fuels are the major contributors to greenhouse gases (GHGs) and other pollutants such as carbon dioxide (CO2), nitrous oxide (NOx), and sulfur dioxide (SO2). The power scheduling problem has, thus, expanded to a multi-objective optimization problem, encompassing environmental emission dispatch (EED). As a result, multi-objective power scheduling (MOPS) seeks to find an optimal dispatch schedule that minimizes both economic costs and several types of environmental emissions.
MOPS is an intricate optimization problem characterized by several complex factors. First, it involves a combinatorial dimensionality explosion, which makes it extraordinarily difficult as the number of generating units increases [4,5]. Each generating unit’s unique characteristics and constraints add another layer of intricacy [6,7]. Ramp rate constraints additionally complicate the decision-making process since they contribute to the non-smooth and discontinuous nature of the input–output characteristics of generating units [8,9,10]. Valve point effects (VPEs) are other factors affecting the input–output characteristics of thermal generating units. Incorporating VPEs into the scheduling problem thereby intensifies the complexities of power scheduling. Furthermore, real-world power systems require multi-period planning to adapt to changing demand patterns and resource availability. Since multi-period planning duplicates the constraints for each period, it compounds the exponential escalation of dimensionality [4,5,11]. Current global environmental policies, such as the Sustainable Development Goals (SDGs), emphasize the reduction of greenhouse gases (GHGs) through integrating renewable energy sources into power networks [12]. While renewables offer environmental benefits, their intermittent nature and weather dependence may lead to significant power supply fluctuations [13]. Additionally, uncertainties in real-world scenarios, including unit and transmission line outages, may complicate power system management even more. Therefore, the MOPS problem poses a multifaceted challenge due to its high dimensionality, multiple constraints spanning multiple periods, non-smooth and non-convex characteristics, and inherent uncertainties.
The MOPS problem has drawn much attention to the development of various optimization models. The methodologies can be grouped into mathematical models, meta-heuristic approaches, and hybrid methods. Mathematical models include priority list (PL), dynamic programming (DP), Lagrangian relaxation (LR), mixed integer linear programming (MILP) [14], and mixed integer quadratic programming (MIQP) [5]. Many of these models are grounded in heuristic principles [4,5]. The ability of these models to guarantee optimal solutions is limited [11], and they face computational inefficiencies, particularly when tackling larger-scale power scheduling problems [15]. Specifically, the PL method is susceptible to poor solution quality [16,17], while DP, MILP, and MIQP encounter dimensionality problems [16,17]. Some specific models also exhibit shortcomings in handling specific constraints, potentially leading to system instability [18]. Moreover, these models are rooted in heuristic principles [4,5], and their ability to guarantee optimal solutions is limited [11]. They also grapple with numerical convergence issues [19]. While effective for scheduling problems with lower dimensions and fewer constraints [20], applying these model-based methods to a more complex scenario raises concerns. Meta-heuristic techniques have been introduced as a promising alternative to address the limitations of mathematical methods. These approaches amalgamate fundamental heuristic principles to enhance the exploration and exploitation of potentially feasible solutions. The most common metaheuristic methods used for power scheduling are differential evolution (DE) [21], genetic algorithms (GA) [22], teaching learning-based optimization (TLBO) [23,24], binary-real-coded genetic algorithm (BRCGA) [25], binary alternative moth-flame optimization (BAMFO) [18], binary bat search algorithm (BBSA) [26], new binary particle swarm optimization (NBPSO) [20], improved dragonfly algorithm particle swarm optimization (iDA-PSO) [27], binary particle swarm optimization (BPSO) [20], quantum-inspired evolutionary algorithm (QEA) [28], quantum evolutionary programming (QEP) [16], quantum inspired binary grey wolf optimizer (QI-BIGWO) [16], quasi-oppositional teaching learning-based algorithm (QOTLBO) [24], binary grey wolf optimizer (BGWO) [29], binary learning particle swarm optimization (BLPSO) [20], binary moth-flame optimization (BMFO) [18], binary coded modified moth flame optimization algorithm (BMMFOA) [30], and teaching learning-based optimization (TLBO) [24]. A comprehensive review of metaheuristic optimization algorithms is provided by [31]. The third category is to use hybrid approaches that combine mathematical and heuristic strategies. The whale optimization algorithm (WOA) [32], the particle swarm optimization (PSO) [33], the grey wolf optimization (GWO) [28,34], the non-dominated sorting genetic algorithm-II (NSGA-II) [35], the non-dominated sorting genetic algorithm-III (NSGA-III) [36], and the evolutionary algorithm based on decomposition (EAD) [37] are some well-known examples applied for power scheduling. Even though these methods integrate diverse strategies, they are still heuristic, and there is no guarantee that they can find global optima [31].
Despite the availability of various optimization techniques, solving MOPS remains an NP-hard (non-deterministic polynomial-time hard) problem [7]. In their attempts to simplify the complexity, many existing studies have tended to overlook the holistic perspective of power scheduling [15]. For instance, several studies [34,38,39] have delved into single-objective power scheduling (SOPS), primarily focusing on minimizing economic costs. While power scheduling is multi-objective, the narrow focus on financial costs might lead to higher emissions, pose health risks, and contribute to climate change and global warming. Furthermore, the studies often involve a limited number of generating units and constraints [40]. Most existing approaches have also focused on single-period problems, simplifying multi-period constraints. Such simplification disregards the broader dynamics and uncertainties that may influence power generation and consumption over a longer planning horizon. Studies like [23,41] have investigated systems with up to 100 units. However, they have overlooked ramp rate constraints and thermal valve point effects (VPEs), which are critical for accurate modeling [8,11]. This means the studies have failed to account for the non-smooth and non-convex nature of the cost and emission functions [8,9,34]. Moreover, effective operational schedules should grapple with power system uncertainties that intermittent renewable energy sources, sudden demand fluctuations, unit outages, and transmission losses may cause. However, many existing optimization models have overlooked these critical aspects of power scheduling [1], potentially leading to suboptimal solutions. These oversights underscore the necessity for a more comprehensive and nuanced approach to addressing the complexities of MOPS.
Due to recent advancements in artificial intelligence (AI), the landscape of decision-making approaches has undergone a paradigm shift. Specifically, reinforcement learning (RL) has emerged as a potent tool for making intelligent decisions in dynamically changing environments [40]. RL, being model-free, is particularly suitable for simulating uncertain real-world scenarios that lack accurate mathematical formulations [29,42]. Its trial-and-error-based approach and inherent adaptability make it a powerful method for handling stochasticity [43]. Despite this potential, RL-based methods are underutilized in power scheduling [42], as shown in Table 1.
Two main RL-based approaches are utilized in the literature for power scheduling: tabular Q-learning [43,44] and function approximation-based Q-learning [5,12,29,40,42,45,46]. The initial tabular Q-learning has been demonstrated in [43], where the pursuit exploration method has been used to examine a four-unit power scheduling problem. Another study utilized tabular Q-learning to solve a three-unit problem. It is reported that Q-learning performed better than the conventional PL method [44]. On the other hand, the function approximation-based Q-learning approach yielded improved solutions over simulated annealing (SA) [47], despite their small-scale experiment involving only eight units. The adoption of the MDP framework in an RL-based approach by [12] was also instrumental in achieving superior outcomes. Furthermore, a centralized Q-learning-based optimization algorithm was compared with a distributed counterpart [29]. Fuzzy Q-learning also outperformed GWO in its application in handling a power system with 10 generators [45]. An adaptive multi-step Q-learning algorithm was employed for a five-unit power system [5], surpassing MIQP. It can be observed that Q-learning methods implemented thus far have primarily been confined to small power systems consisting of 3 to 10 units. This limitation arises from the challenges posed by extensive state and action spaces due to the inherently combinatorial nature of the problem. The studies by [40,46] have addressed up to 30-unit power systems and reported better performance than MILP. Though up to 100 units are implemented by [42] and the solutions are better than GA, the study compromised ramp rate constraints and VPEs.
Table 1. Summary of RL-based power scheduling studies.
Table 1. Summary of RL-based power scheduling studies.
ReferenceObjectiveUnitsConstraintsVPEs
Production CapacityOperating DurationRamp RatesReserve
[43]Single4YesNoNoNoNo
[44]Single3YesNoNoNoNo
[47]Single8YesYesNoNoNo
[12]Single10YesYesNoNoNo
[45]Single10YesYesYesYesNo
[29]Single 10YesNoNoNoYes
[5]Single5YesYesYesNoNo
[40]Single 30YesYesNoNoNo
[46]Single 30YesYesNoNoNo
[42]Single 100YesYesNoYesNo
Generally, the existing RL-based studies have demonstrated their superiority over traditional methods such as PL, SA, LR, MILP, MIQP, GWO, and GA. However, most of these achievements have been confined to smaller-scale problems. Additionally, most RL models have neglected to ensure reserve capacity to handle unforeseen circumstances like sudden spikes in demand, generating unit failures, or transmission line losses. Further, the impact of ramp rate constraints and VPEs has frequently been omitted from consideration in the models. It is also noted that the previous studies focused on the single-objective problem of minimizing economic costs. Thus, the existing studies inadvertently neglected the crucial aspect of the environmental impact of carbon emissions. More specifically, no known RL-based method has been applied to MOPS problems. Solutions derived from simplified power scheduling problems may not accurately capture the complexities of real-world power systems [42].
These existing disparities underscore a significant research gap that emphasizes the pressing need to develop a scalable RL algorithm that could potentially unlock innovative solutions and pave the way for more effective and reliable power scheduling techniques without compromising the characteristics of a realistic power system. This study introduces an innovative MOPS optimization approach based on multi-agent reinforcement learning. The proposed method utilizes a contextually adaptive multi-agent simulation environment where power-generating units are represented as agents. The dynamics of MOPS are simulated within this environment using Markov decision processes (MDPs), which are then used to train a multi-agent deep RL (MADRL) model. The solution for the optimization problem is derived from the trained MADRL model.
In summary, the contributions of this article are as follows:
  • The primary contribution is the introduction of a pioneering MADRL-based optimization algorithm that can solve single- to tri-objective power scheduling problems. The algorithm harnesses the power of MADRL to decisively confront the intricate challenges of MOPS.
  • Unlike existing methods, the proposed algorithm does not confine itself to a limited planning horizon and a fixed number of generating units. It also encompasses a comprehensive set of unit-specific (including ramp rates and VPEs) and system-level constraints such as reserve availability.
  • Another distinctive contribution to our approach is developing a contextually adaptive multi-agent simulation environment. The environment is used to decompose the MOPS problem into sequential MDPs. It can contextually correct agents’ illegal decisions and adjust excess and shortages of supply capacities. By accelerating agents’ learning and reducing model training time, the simulation environment may significantly enhance the efficiency of the entire optimization process.
  • The simulation environment is not specifically tailored to train a specific RL model but is model agnostic. This adaptability allows researchers and practitioners to train and explore diverse types of RL models for solving power scheduling.
  • Unlike traditional models with exponential dimensionality (i.e., O ( 2 n ) for n generating units), the proposed algorithm has linear dimensionality (i.e., O ( 2 n ) ). This characteristic underscores its scalability and better performance to handle large-scale problems compared to existing methods.
  • The algorithm’s programming code has been verified by Code Ocean for quality and computational reproducibility (https://doi.org/10.24433/CO.9235622.v1) and published as an open-source software package [48]. This initiative may facilitate results replication and foster a spirit of experimentation and further extensions within the research community.
The remainder of this article is organized as follows: Section 2 describes the description and formulation of the MOPS objective problem. Section 3 briefly outlines the procedural framework of the proposed optimization methodology. Section 4 presents practical applications and results, and Section 5 presents the concluding remarks.

2. MOPS Problem Formulation

Consider a power scheduling problem consisting of n thermal generating units, all subject to optimization, over a planning horizon with T periods. Given operating durations ( t t i R ; i ) at each period t , two decision variables are associated with each unit. The first is a unit commitment (UC), indicating whether a unit is committed t t i > 0 or not t t i < 0 , and the second is the power output p t i , t , i .

2.1. Cost Objective Function

The operating cost ($/period) of each unit i during each period t encompasses three components: production cost C o n p t i , startup cost C t i s u , and shutdown cost c t i s d [49]. Considering operating durations and transitions between online and offline statuses, its comprehensive computation is specified as follows:
C t i = t t i > 0 C o n p t i + t t i > 0 t t 1 , i < 0 C t i s u + t t i < 0 t t 1 , i > 0 C t i s d ; t , i .
The production cost C o n p t i is typically represented by a quadratic function of the power output p t i . Unlike the usual model-based optimization methods, the proposed MADRL approach does not necessitate approximating production costs with smooth and convex functions that disregard VPEs. Consequently, the production cost function incorporates a sinusoidal term to account for VPEs [50], which can be written as follows:
C o n ( p t i ) = α i c p t i 2 + β i c p t i + δ i c + | ρ i c s i n [ φ i c ( p i m i n p t i ) ] | ; t , i .
The objective of ECD is to minimize the total operating costs over the entire planning horizon, defined as follows:
C = t = 1 T C t = t = 1 T i = 1 n C t i .

2.2. Emission Objective Function

As mentioned earlier, electricity generation is the major contributor to GHGs and environmental pollutants. Specifically, the combustion of carbon-containing fossil fuels, such as coal, oil, or natural gas, results in the production of CO2. CO2 is a major contributor to climate change and/or global warming. The combustion of fuels containing sulfur compounds also releases SO2. SO2 is a precursor to acid rain and can have detrimental effects on the quality of the air. Consequently, beyond the economic implications, addressing environmental emissions becomes crucial to mitigating the impacts of global warming and air pollution.
Similar to the costs associated with power generation, emissions are not directly tied to power outputs and exhibit variations among different generating units. Mathematically, the operating emissions (lbs/hour) for each unit i during each period t are the sum of startup E t i s u , shutdown E t i s d , and production E o n p t i emissions.
E t i = t t i > 0 E o n p t i + t t i > 0 t t 1 , i < 0 E t i s u + t t i < 0 t t 1 , i > 0 E t i s d ; t , i
Like the production cost functions, non-smooth and non-convex functions are used for the production emissions, which are expressed as follows:
E o n p t i = α i e p t i 2 + β i e p t i + δ i e + ρ i e exp φ i e p t i ; t , i .
The total daily emissions over the entire planning horizon can then be consolidated as follows:
E = t = 1 T E t = t = 1 T i = 1 n E t i .
The costs and emissions linked to generating units exhibit an inherent conflict [51]. Hence, operating at minimum cost alone (i.e., Equation (3)) or at absolute emission level (i.e., Equation (6)) is no longer acceptable because minimizing one leads to an increase in the other.

2.3. MOPS Objective Function

The objective function for MOPS is usually formulated by combining the separate cost and emission functions in Equations (3) and (6), respectively, into one. Traditionally, various methods have been widely employed, such as emission constraints, weighted-sum approaches, and cost-penalty factors. However, these methods come with inherent limitations. Emission constraints often fall short of adapting to real-time changes and tend to prioritize compliance over effective emission reduction strategies [42]. The weighted-sum approach faces challenges in striking the right trade-offs between costs and emissions [31]. The specific value of the weight might have an insignificant impact when costs and emissions are quite different in size [52]. It also exhibits limited efficiency for non-convex Pareto-optimal fronts [53] and applies only to convex cost and emission functions [54]. The cost-penalty method is used to convert emission curves into equivalent cost curves. However, its application is limited to increasing cost and emission functions. Additionally, its estimates are often unrealistically small or large. Consequently, this approach fails to capture the intricate relationships among the components [29]. Merging conflicting objectives may further obscure the trade-off relationship between cost and emissions, resulting in suboptimal solutions that do not fully exploit the potential for minimizing costs and emissions independently. This hindrance makes identifying Pareto-optimal solutions representing the best compromises between conflicting objectives challenging. In addressing these challenges, this study proposes a new hybrid approach to formulating the objective function. First, the production cost function in Equation (2) and the production emission function in Equation (5) are unified using both weight hyperparameters and unit-specific cost-to-emission conversion parameters, as follows:
Φ o n p t 1 , p t 2 , p t n = ω i = 1 n t t i > 0 C t i o n p t i + 1 ω i = 1 n t t i > 0 η i E t i o n p t i
where ω and η i denote the weight hyperparameter ( 0 < ω < 1 ) and the cost-to-emission conversion parameter ( η i > 0 ), respectively. Equation (7) represents a bi-objective power scheduling problem, holding economic costs on one side and emissions on the other side. Considering m types of environmental emissions, it can be extended to a general MOPS problem, and written as follows:
Φ o n p t 1 , p t 2 , p t n = i = 1 n t t i > 0 ω 0 C t i o n p t i + k = 1 m t t i > 0 ω k η i k E t i k o n p t i
where ω k represents the weight associated with objective k such that ω k 0 and k = 0 m ω k = 1 . The parameter η i k is the cost-to-emission conversion factor for unit i corresponding to emission type k . This formulation not only allows the representation of emissions as continuous variables but also maintains their relative importance within the integrated production function.
Second, the startup and shutdown costs and emissions are aggregated separately. Consequently, the objective function of the MOPS consists of three separate components, which can be expressed as follows:
Φ = t = 1 T Φ o n p t 1 , p t 2 , p t n ,   t = 1 T C t s u , s d ,   t = 1 T E t s u , s d
where C t s u , s d = i = 1 n t t i > 0 1 t t 1 , i < 0 C t i s u + 1 t t i > 0 t t 1 , i < 0 C t i s d ; t and E t s u , s d = i = 1 n t t i > 0 1 t t 1 , i < 0 E t i s u + 1 t t i > 0 t t 1 , i < 0 E t i s d ; t .
In contrast to existing approaches, the objective function in Equation (9) is not intrinsically converted to a single-objective function. Rather, it consists of three components: the transformed production value, the cost of startup and shutdown, and the emissions during startup and shutdown. This innovative formulation is designed to tackle the limitations of traditional methods, offering a more resilient and flexible solution to the simultaneous optimization of cost and emissions. For the sake of simplicity, let us denote each objective by O k k = 0,1 , . . m . The operating cost function in Equation (1) and the operating emission function in Equation (5) specific to each objective can be written as follows:
O t i k = t t i > 0 O k o n p t i + t t i > 0 t t 1 , i < 0 O t i k s u + t t i < 0 t t 1 , i > 0 O t i k s d
where O k o n p t i , O t i k s u , and O t i k s d are the production, startup, and shutdown values corresponding to objective k , respectively.

2.3.1. Constraints

The MOPS objective function in Equation (9) is optimized subject to Equations (11)–(15).
Power production capacities:
t t i > 0 p i m i n p t i t t i > 0 p i m a x ; t , i
Maximum ramp rates:
t t 1 , i > 0 t t i > 0 p t 1 , i p i d o w n p t i t t 1 , i > 0 t t i > 0 p t 1 , i + p i d o w n ; t , i
Minimum operating durations:
t t i O N t i u p ; t , i   and   t t i O F F t i d o w n ; t , i
Supply and demand balance:
i = 1 n t t i > 0 p t i = d t ; t
Minimum reserve constraint:
i = 1 n t t i > 0 p t i m a x 1 + r d t ; t

2.3.2. Cost-to-Emission Conversion Factors

The MOPS objective function in Equation (9) requires estimating the cost-to-emission scale for each unit, η i ($/lbs). For this purpose, a custom function is introduced using finite difference gradients as follows:
η i k = exp C o n p i / E k o n p i max C o n p i E k o n p i ; i min C o n p i E k o n p i ; i
where C o n p i = C o n p i m a x C o n p i m i n , and E k o n p i = E k o n p i m a x E k o n p i m i n ; i , where k = 1,2 , . . m . Over- or underestimations are mitigated by standardized gradients, and the exponential function ensures that the estimates remain non-negative.

2.3.3. Sensitivity Analyses for Weights

Unlike single-objective optimization, there is no unique optimal solution to the MOPS problem. Instead, the results lead to identifying compromised or non-dominated (i.e., Pareto) optimal solutions [51]. To address this complexity, agents representing various generating units can be assumed to have imprecise goals. Fuzzy logic theory can then be applied to discern the optimal trade-off. To do so, the Pareto-optimal solutions for different weight combinations should first be stored in a predefined repository. Next, a fuzzy membership value μ O k ω is calculated for each objective function as follows:
μ O k ω = 1 , O k ω O k b e s t O k w o r s t O k ω O k w o r s t O k b e s t , O k b e s t < O k ω < O k w o r s t 0 , O k ω O k w o r s t ; ω w ; k = 0,1 , , m
where O k b e s t is the single objective best value, O k w o r s t is its corresponding worst value, and w represents the set of all stored weight combinations, w = ω ; ω 0,1 m + 1 . The performance of each non-dominated solution can then be measured in terms of cardinal priority. Cardinal priority provides a normalized membership function across all objectives and non-dominated solutions. It can be determined as follows:
μ O ω = k = 0 m μ O k ω / ω w k = 0 m μ O k ω .
The weight combination that achieves the maximum cardinal priority ω a r g m a x μ O ω , ω w can be selected as the optimal trade-off.

3. Proposed Methodology

This study’s methodology revolves around multi-agent reinforcement learning, an AI paradigm where multiple agents learn through trial and error within a dynamic environment. Individual power-generating units are represented as agents within a custom multi-agent simulation environment. The agents can independently determine their commitment statuses and load dispatches as long as the unit-specific constraints in Equations (11)–(13) are met. As a result, the designation of the multi-agent simulation environment aligns with the principles of agent-based simulation. Unlike traditional agent-based simulation, the agents are not entirely autonomous but guided by specific goals. The goals correspond to the system-level constraints specified in Equations (14) and (15), which significantly shape the joint action of all agents. Specifically, the decisions of all agents must meet two crucial conditions. First, the total power supply from all ON agents must fulfill the demand, including reserve capacity, at each period of the planning horizon. Second, the agents collaboratively strive to minimize the value of the MOPS objective function across the entire planning horizon, which is specified in Equation (9). As a result, the interactions among agents within the multi-agent simulation environment follow a cooperative multi-agent reinforcement learning approach [55].
Considering an hourly divided day as a power scheduling horizon, each hour serves as a timestep t , and the entire planning horizon constitutes an episode. At each timestep t of an episode, there are four fundamental components in the RL framework: state space S = s , action space A = a , transition (probability) function P , and a scalar reward R . The formalization of each element for the MOPS problem is as follows:
State Space ( S ): For each timestep t of an episode, the system will be in state s t 1 = t , p t 1 m i n , p t 1 m a x , t t 1 , d t R 1 × 3 n + 2 , where p t 1 m i n = p t 1,1 m i n , p t 1,2 m i n , , p t 1 , n m i n and p t 1 m a x = p t 1,1 m a x , p t 1,2 m a x , , p t 1 , n m a x are the vectors of minimum and maximum capacities, respectively, t t 1 = t t 1,1 , t t 1,2 , , t t 1 , n is the operating durations, and d t is the demand to be satisfied. If ramp rates are not considered, the state may be simplified to s t 1 = t , t t 1 , d t R 1 × n + 2 . The state space for each timestep can be described as S = s t 1 .
Action Space ( A ): Each of the n agents will have two decisions (OFF or ON) in each state s t 1 . The action space A will consist of 2 n unit commitments (i.e., A = 0,1 2 n × n ) for each timestep t . The decisions of all agents will then constitute an n -dimensional action a t 1 = a t 1,1 , a t 1,2 , , a t 1 , n = 0,1 n A .
Transition Function  P : After agents take decision a t 1 A in state s t 1 S , the transition (or probability) function P s t | s t 1 , a t 1 leads to the next state s t . Adhering to all the constraints in Equations (11)–(15), the transition depends solely on state s t 1 and action a t 1 . If any of the constraints is violated, it would not be legitimate to advance to the succeeding state s t .
Reward function ( R ): Reward is a reinforcement signal measuring the performance of agents’ decisions in transitioning to the next state s t S . It is a predefined function of the action a t 1 A and state s t 1 S , that is, r t = R s t 1 , a t 1 , s t .

3.1. Multi-Agent Simulation Environment

The key elements of MARL satisfy the properties of an MDP [55]. Consequently, the MOPS dynamics can also be formulated as a 4-tuple S , A , P , R MDP that will be simulated within the multi-agent simulation environment, as described below.
[Step 0] Inputs: The environment should be provided with parameters of power-generating units, demand profile, weight hyperparameters, and other specifications (such as the type of optimization: single, bi-objective, or tri-objective).
[Step 1] Initialization: The environment is initialized with state as s 0 = 0 , p 0 m i n , p 0 m a x , t 0 , d 1 .
[Step 2] Constraints: Units failing to meet their minimum down time ( t i d o w n ) are kept OFF ( u t i 0 = 1). Similarly, units not fulfilling their minimum up time ( t i u p ) are constrained to be ON ( u t i 1 = 1). These constraints can be defined as follows:
u t i 0 = 1   if t i d o w n < t t 1 , i < 0 ;   u t i 1 = 1   if   0 < t t 1 , i < t i u p ;   i .
[Step 3] Startup and shutdown values: Startup cost is determined based on the duration during which a unit has been continuously offline (OFF) until a certain timestep t . This cost is commonly approximated using a staircase function given as follows:
C t i s u = C i h o t ,   i f t i d o w n + t i c o l d t t 1 , i t i d o w n C i c o l d ,   i f   t t 1 , i < t i d o w n + t i c o l d                                               ; i .
On the other hand, startup emissions E t i s u , shutdown costs C t i s u , and shutdown emissions E t i s d are all assumed to be zero as usual in the existing literature.
[Step 4] Transition { P s t | s t 1 , a t 1 } : At each state s t 1 , the transition function will receive the decision a t 1 of all n agents as an input. The transition function plays a critical role in ensuring that the MARL framework aligns with the constraints of the MOPS problem. Accordingly, the legality of each agent’s decision is verified first based on the constraints determined in Equation (19). The collective decisions are subsequently assessed to confirm if the total supply capacity can meet the demand required, including reserve, for that timestep. Details of these adjustments are explained below.
[Step 4.1] Legalize agents’ decisions: Upon receiving the action a t 1 of agents, the transition function will scrutinize the legality of each agent’s decision a t 1 , i a t 1 . Any agent’s decision conflicting with Equation (19) is corrected as follows:
a t 1 , i = 1   if   a t 1 , i = 0   |   u t i 1 = 1 ;   a t 1 , i = 0   if   a t 1 , i = 1   |   u t i 0 = 0 ;   i .
Equation (21) represents the appropriate adjustments to be made if there are ON units that should have been OFF, or vice versa.
[Step 4.2] Priority list: To facilitate capacity adjustments, priority list approaches [38,39,40] are usually adopted. Existing priority list methods are limited to increasing cost and emission functions. These methods additionally disregard startup and shutdown impacts. To address these limitations, this study has made some adjustments. First, the minimum marginal production values per MW of the individual objectives are modified as follows:
λ i k m i n = 1 p i m a x O k o n p i ; i , k   where   p i = p i m a x ,     i f   O k o n p i m a x   O k o n p i m i n p i m i n ,     i f   O k o n p i m a x < O k o n p i m i n ; i , k .
Equation (22) accounts for the trends of cost and emissions functions, but it does not incorporate units’ startup and shutdown impacts. By incorporating startup and shutdown impacts together with the minimum up- and down-time durations, Equation (23) determines the minimum marginal operating values for each objective.
Λ i k m i n = λ i k m i n + 1 p i m a x a t 1 , i t t 1 , i < 0 O t i k s u t i u p + 1 a t 1 , i t t 1 , i > 0 O t i k s d t i d o w n ; i , k .
The priority list corresponding to the MOPS objective function can be obtained by averaging the priority values of the individual objectives, defined as follows:
Λ i m i n = 1 m + 1 k = 0 m Λ i k m i n ; i .
[Step 4.3] Capacity adjustments: Based on the order of values in Equation (24), the simulation environment orchestrates capacity adjustments to ensure a reliable power supply that satisfies current and future demands. First, there may be cases where future demands d t * ; t < t * < 24 cannot be met during the downtime of some units. Those units must remain online, even if they have fulfilled their minimum uptime durations. Second, the simulation environment will be able to manage both excess and shortage of power supply capacities as follows:
  • If there is a shortage of capacity (i.e., i = 1 n a t 1 , i p t 1 , i m a x < 1 + r d t ), unconstrained OFF units are turned ON ( a t 1 , i = 1 | u t i 1 = 0 ) in the increasing order of Λ i m i n until the demand, including reserve, is met (i.e., i = 1 n a t 1 , i p t 1 , i m a x 1 + r d t ) or no unconstrained OFF unit remains.
  • If there is excess capacity (i.e., i = 1 n a t 1 , i p t 1 , i m i n > 1 + r d t ), unconstrained ON units are turned OFF ( a t 1 , i = 0 | u t 1 , i 0 = 1 ) in the decreasing order of Λ i m i n until supply matches demand, including reserve (i.e., i = 1 n a t 1 , i p t 1 , i m i n 1 + r d t ) or no unconstrained ON units are left.
[Step 4.4] Optimal loads of MOPS production function: The optimal loads in the unified production function Φ o n p t i defined in Equation (7) can be determined using sequential least squares programming (SLSP) [56]. These loads p t i , i can be determined as follows:
p t = a r g m i n p t 1 , p t 2 , , p t n Φ o n p t 1 , p t 2 , p t n .
[Step 4.5] Incomplete episode and terminal state: Complete episodes share the same terminal state s 23 + , possibly with different state values. However, determining unit commitments and load dispatches for all timesteps is not always feasible. Despite the contextual correction capabilities of the simulation environment, certain scenarios may pose challenges. There might not be enough unconstrained ON units to switch OFF to adjust excess capacity. Conversely, there might not be enough unconstrained OFF units to switch ON during a supply shortage. In such situations, transitioning from the current state s t 1 to the next succeeding state s t is not possible, and this state s t 1 is also designated as a terminal state (i.e., s t 1 + ). This renders the corresponding episode incomplete. In other words, an episode is incomplete if s t 1 + = 1 whenever t < 24 . It should be noted that the terminal state varies from episode to episode, resulting in the number of timesteps across episodes being a random variable [55].
[Step 4.6] Total operating values of individual objective functions: The total operating value O t k of objective k is calculated by summing its startup, shutdown, and production values, which can be written as follows:
O t k = i = 1 n a t 1 , i   O k o n p t i + a t 1 , i t t 1 , i < 0 O t i k s u + a t 1 , i t t i > 0 O t i k s d , k
where the production value O k o n p t i is evaluated at the optimal loads determined in Equation (25). In episodic decision-making scenarios, it is usually needed to prevent incomplete episodes. To ensure this, large penalties are often recommended [5]. In this study, the penalty imposed ensures that episodes with fewer timesteps incur higher operating values, underscoring the significance of complete episodes for effective learning and decision-making. This penalty can be defined as follows:
O t k = i = 1 n O k o n p i m a x + t 1 23 O k o n , m a x i = 1 n O k o n p i m a x ; k
where O k o n p i m a x represents the maximum capacity production value and O k o n , m a x denotes the maximum possible production value, which is computed as O k o n , m a x = i = 1 n λ i k m a x p i m a x . Here, λ i k m a x is given as follows:
λ i k m a x = 1 p i m i n O k o n p i   where   p i = p i m i n ,     i f   p i m i n β 2 α β 2 α ,       i f   p i m i n < β 2 α ; i .
Equation (28) is like Equation (22), but represents the maximum marginal production value per MW, considering the trends of the cost and emission functions. It is crucial to note that the actual operating cost in Equation (26) is utilized whenever s t 1 + = 0 , while the penalty in Equation (27) is applied when the episode becomes incomplete (i.e., when s t 1 + = 1 for t < 24 ).
[Step 4.7] Determine the next state: Depending on the contextually corrected action a t 1 in state s t 1 , the simulation environment is initialized, or the subsequent state s t is obtained through updating s t 1 . Details of these scenarios are described in [42].
[Step 4.8] Reward function: The reward function r t = R s t 1 , a t 1 , s t is defined as the inversed average of the normalized values of all individual objectives. It can be represented as follows:
r t = 1 m + 1 k = 0 m O t k O k m i n O k m a x O k m i n 1
where O k m i n represents the minimum possible production value, which is defined as O k m i n = min λ i k m i n p i m i n ; i T .
[Step 5] Return: Executing action a t 1 in state s t 1 in the transition function will mainly return a three-tuple ( s t , r t , s t 1 + ), where s t represents the next state, r t represents the reward defined in Equation (29), and s t 1 + is an indicator whether the current state is terminal. Additionally, the output includes unit commitments, optimal loads, and other related information.
[Step 6] Proceed to the next timestep or reset the environment: If s t 1 + = 0, the simulation continues to the next timestep of the episode, executing another action of agents in Step 4. However, if s t 1 + = 1, the simulation environment will be reset to its initial status (proceeding to Step 1) to start the next new episode. This ensures a clear distinction in the simulation process based on the completeness of the episodes.
Figure 1 summarizes the process of simulating the power scheduling dynamics in the form of MDPs, and the pseudocode is presented in Algorithm 1.
Algorithm 1. Pseudocode for simulating power scheduling as MDPs
Systems 12 00106 i001

3.2. Multi-Agent Deep Q-Network (MADQN)

In the framework of RL, the transition function P s t , r t s t 1 , a t 1 defines the probability distribution of the next state s t with reward r t for taking an action a t 1 A s t 1 in state s t 1 S . This mapping is called a policy π a t 1 s t 1 [55], which is a guide that helps the agents choose action a t 1 given the state s t 1 at each timestep t of an episode. The goal of the different cooperative agents is to learn a policy π a t 1 s t 1 that maximizes their long-run cumulative return (sum of rewards), determined as G t = t = 1 T γ   t 1 r t , where γ is a discount rate 0 γ 1 . The expected return of the state V π s following a policy π is defined as V π s = E π G t s t 1 = s . Similarly, the action-value function Q π s , a , representing the expected return of taking an action a t 1 in state s t 1 , is given by:
Q π s , a = E π G t s t 1 = s , a t 1 = a .
The optimal policy is now denoted as π s | a and is defined as π s | a = a r g m a x π V π s for all s S or π s | a = a r g m a x π Q π s , a for all a A s . As a result, the action value corresponding to the optimal policy is as follows:
Q π * s , a = ma x π Q π s , a ; s , a S × A
where Q π s , a represents the action value defined in Equation (30). However, calculating exact values as defined in Equation (31) for all state-action pairs s , a is infeasible due to the exponential growth of the state-action pairs. Instead, these values can be estimated through function approximation methods [29]. This study employs a deep Q-network (DQN) to approximate the action value function. DQN is a robust technique as it does not require a precise mathematical definition of a loss function and can identify optimal solutions, even for nonconvex loss functions [29]. The specifics of the DQN model, including its network architecture, exploration strategy, loss function, and unique features, are detailed in [48].

3.2.1. Model Architecture

The DQN model employed in this study utilizes a single-hidden-layer feedforward neural network model with ReLU activation functions in both the hidden and output layers. The input of the network comprises the MDP dynamics represented by s t 1 = t , p t 1 m i n , p t 1 m a x , t t 1 , d t , which are simulated within the multi-agent simulation environment. The network output represents the action value function Q π s , a | w , where w consists of the model’s weights. The model is designed to predict two values for each agent. In this setup, each pair of nodes corresponds to the two possible decisions (i.e., ON and OFF) for each of the n agents. This approach is necessitated by the complexity of the action space A , which inherently poses a formidable challenge due to its exponential growth concerning the number of agents involved (resulting in 2 n possible unit commitments). To mitigate this complexity, the network is designed with 2 n output nodes, instead of parameterizing it into 2 n output nodes. This streamlined network parametrization is a strategic solution to handle the computational constraints posed by the exponential growth in the action space. This simplified network setup is strategic for managing computational complexity while ensuring an efficient modeling process. Importantly, the network design is decentralized, where each agent has a dedicated pair of output nodes. As a result, the DQN is referred to as the MADQN. This approach results in the network output becoming a linear function of the number of agents, enabling a more manageable and practical implementation within the multi-agent system.
In this framework, an ϵ -greedy exploration approach is used. Accordingly, agents autonomously make their optimal decision a t 1 , i ; t within state s t 1 with probability 1 ϵ . Conversely, with probability ϵ , each agent explores the environment by deciding randomly. In both conditions, once the action a t 1 of all agents is formed, it serves as an input to the transition function in the simulation environment, initiating the next state s t , which will be the input for the MADQN model.
It is noteworthy that while MADQN is used in this study, the multi-agent simulation environment is not specifically designed for a particular RL model. It remains adaptable, allowing for the training of various RL models based on user preferences.

3.2.2. Experience Relay

In the context of power scheduling problems, an action taken in a particular state affects the immediate reward and has repercussions on a sequence of future states. Such interdependence among the successive MDPs poses a challenge when applying the standard MADQN [5]. The MADQN can be enhanced by incorporating an experience replay to address this issue. Experience replay involves storing the transition tuples in a replay buffer, denoted as B = s t 1 , a t 1 , r t , s t , and then using a random batch of experiences s , a , r , s to estimate the action values. This approach helps break the temporal correlations among the MDPs [57]. By learning from random past experiences, the agents can more effectively explore the state-action space in diverse situations. This approach enhances the learning process by mitigating the challenges of autocorrelated sequential MDPs.

3.2.3. Loss Function and Parameter Updates

The choice of an appropriate loss function is crucial in RL to ensure stable and effective training of the model. To calculate the loss, MADQN incorporates target and main networks [45]. The target network generates target action values and is updated periodically from the primary network Q s , a w . Its values are computed as r + γ   max a Q π s , a w , where w is the weight of the target network. Due to the diverse variables present in the state input for the MADQN model, large errors in action-value estimates can occur, which may potentially lead to unstable training. In such situations, the Huber loss function proves to be more appropriate [55], which is expressed as:
J w = E 0.5 r + γ   max a Q s , a w Q s , a w 2 ,   i f   r + γ   max a Q s , a w Q s , a w < 1 r + γ   max a Q s , a w Q s , a w 0.5 ,   i f   r + γ   max a Q s , a w Q s , a w 1
where s , a , r , s represent a batch of experiences sampled from the replay memory B . The update rule for the network parameters is also expressed as:
w w + ξ r + γ   max a Q s , a w Q s , a w Q s , a w
where ξ is a learning rate. Huber loss can effectively handle different error values, ensuring stable learning [55]. It becomes an absolute error (i.e., L1 loss) for larger errors and resembles a mean-squared error (i.e., L2 loss) for small errors. This property of Huber loss enhances the model’s resilience to noisy data and outliers, leading to more reliable and stable training outcomes in the MADQN framework.
Figure 2 demonstrates how the multi-agent simulation environment and MADQN interact, combining the replay memory and loss function.

4. Practical Applications

The proposed MARL optimization algorithm addresses the MOPS problem with varying units and constraints. The implementation leverages the PyTorch deep learning framework, utilizing Python 3.10. It was executed on a desktop computer with an 11th Gen Intel (R) Core (TM) i7-11700 processor operating at 2.50 GHz, 16 GB of RAM, and eight parallel workers. In the write-up of the manuscript, while QuillBot has been used for language-related aspects, the substantive content and findings resulted from human intellectual effort.

4.1. Experimental Settings

4.1.1. Specifications of Generating Units

Profiles of the generating units, including the cost, emission, and ramp rate parameters, are taken from [54]. The VPE parameters of the cost functions are obtained from [58], and those of the emission functions are accessed from [21]. Moreover, the parameters specific to CO2 and SO2 for the tri-objective problems are drawn from [36]. Operational constraints, including generation capacity, minimum operating time durations, and a 10% reserve, remain the same across all test systems. Notably, shutdown costs and startup emissions are typically assumed to be negligible compared with other expenses [59], and their specific values are absent in the existing literature. Consequently, these parameters are uniformly set to zero in all the experimental analyses, although their integration into the algorithm’s framework has been duly considered. The proposed algorithm is not necessarily confined to a fixed timeframe. Despite this, all the experimental analyses are conducted over a 24 h planning horizon, with demand data from [54].

4.1.2. Parameters and Hyperparameters

The cost-to-emission conversion parameters significantly impact the optimization trade-offs between cost and emission objectives. Estimates of these parameters are presented in Table 2, which are calculated using Equation (16). The other crucial factors affecting MOPS optimization are the weight hyperparameters assigned to the different objectives. These weights are determined before fine-tuning the hyperparameters of the neural network model. As previously noted, the MADQN model utilized a single-hidden-layer neural network incorporating ReLU activation functions in both the hidden and output layers. With a preliminary configuration of the model, a thorough sensitivity analysis is conducted using fuzzy set theory. First, each objective function is individually optimized. Next, a set of 100 evenly distributed weights, ranging from 0 to 1, is considered for the bi-objective problems. Then, a membership function is created for each objective function using Equation (17), and the cardinality values are determined using Equation (18). The results of the top 10 weights with the largest cardinality values for the first 4 test systems are presented in Table 3. In the context of tri-objective scheduling problems, a set of 200 random combinations of weights is sampled from the Dirichlet probability distribution. Table 4 provides the top 10 weight combinations, each possessing the highest cardinality values.

4.1.3. MADQN Model Configurations

Following the selection of promising weights for the different objectives, the hyperparameters of the MADQN are tuned. In the network configuration, generation capacities were initially assumed to be included as part of the model’s input when ramp rate constraints were considered. However, integrating these generation capacities into state input destabilizes training stability. Consequently, the input layer in all the test system models consists of n + 2 nodes, encompassing three elements: the timestep of the planning horizon, the operating time durations, and the demand requirement to be satisfied. All the test systems employed a uniform learning rate of 0.01 and a consistent discount factor of 0.99. The optimization has been performed with the Adam optimizer and the Huber loss function. Furthermore, the exploration–exploitation strategy involves a maximum exploration rate set at 1 and a minimum rate of 0. An exponential decay rate associated with the number of training episodes determines the gradual reduction in exploration as the model’s training progresses. As the model employs experience replay, the batch size is consistently set to match the size of replay memory. The uniformity ensures a consistent approach to learning and updating the model parameters across different scenarios. Table 5 shows the parameters and optimal hyperparameters of the MADQN models for the different test systems.

4.2. Comparative Results

4.2.1. Test System I: Bi-Objective Problem without Ramp Rate Constraints and No VPEs

Test system I comprises 10 generating units, each with smooth and convex cost and emission functions. The unit commitments, loads, and actual reserve percentages resulting from the proposed MADRL algorithm are presented in Table 6. The results of the proposed MADRL approach are compared with those of [23,34]. The former employed TLBO and achieved a daily operating cost of $578,445.5 with an emission of 37,302.9 lbs, while the proposed MADRL results in an operating cost of $563,990.6 and an emission level of 41,310.4 lbs. If economic considerations, which are critical in various industrial and business applications, are paramount, MADRL may be preferred. On the other hand, if environmental concerns take precedence, TLBO might be considered. As a multi-objective optimization problem, MADRL strikes a reasonable balance between operating costs and environmental emissions. Hence, the higher daily operating cost for a slightly lower emission introduced by TLBO makes its solutions less viable from an economic standpoint.
Conversely, RCGWO was adopted in [34], which reported a daily operating cost of $568,655.3 and an emission of 43,759.7 lbs. Table 7 presents a detailed comparative analysis of outcomes obtained by the proposed MADRL methodology and the RCGWO [34] method. The hourly costs and emissions of both methods are also visualized in Figure 3. The slightly higher startup cost with the MADRL algorithm indicates an increase in the number of committed units. However, the efficiency of MADRL is demonstrated with a lower daily production cost of $563,990.6 compared with the cost of $565,115.3 using RCGWO. This economic efficiency positions MADRL as a promising solution. Furthermore, MADRL established a new environmental benchmark, emitting only 41,310.4 lbs of pollutants compared with RCGWO’s 43,759.7 lbs. Particularly noteworthy is the substantial reduction in hourly emissions, emphasizing the environmental conscientiousness of MADRL alongside its economic prowess. Excelling in both economic viability and environmental responsibility, the results underscore the holistic efficiency of MADRL.

4.2.2. Test System II: Ramp Rates Constrained Bi-Objective Problem without VPEs

Test system II is like test system I, except it is ramp rate constrained. No additional units are committed compared to the optimal schedule without ramp rate constraints. However, the dispatches of online generating units are strategically altered due to the inclusion of ramp rate constraints, which leads to changes in both operating costs and emissions. The results of the proposed MADRL and RCGWO [34] are shown in Table 8 and visually presented in Figure 4. With respect to startup costs, RCGWO [34] has a marginal advantage, requiring only $3540.0 compared with the slightly higher investment of $4090.0 of MADRL. However, the real distinction lies in production costs, where MADRL excels at a daily expenditure of $563,106.8 compared with $565,183.9 from RCGWO. Considering both the startup and production costs, MADRL demonstrates its economic efficiency, with a total daily cost of $567,196.8 compared with the total cost of $568,723.9 using RCGWO. Moreover, the proposed MADRL stands out as an environmentally conscious method, emitting only 41,810.8 lbs of pollutants per day, notably lower than the 43,733.6 lbs of RCGWO. These comparative results reveal the superior economic and environmental performance of the MADRL method over the RCGWO approach. A substantial decrease in production expenditure without compromising ecological standards underscores the efficiency and viability of the proposed MADRL approach. The balanced approach, evident in lowered emissions and reduced costs, indicates the potential contribution of MADRL to a greener, sustainable energy future. Moreover, Figure 4 shows significant hourly emission reductions similar to the hourly emissions for the test system without ramp rates. This consistent performance of MADRL across the entire planning horizon underscores its robustness and reliability.

4.2.3. Test System III: Bi-Objective Problem with VPEs and No Ramp Rates

The scenario associated with VPEs is found in test system III (results are not presented). Even though there is no direct comparison with existing literature incorporating VPEs, the potential of MADRL can be appreciated through the numerical results. The total daily cost with VPEs rises to $591,918.6 compared with $568,080.6 without VPEs. Similarly, the total daily emissions increased to 43,796.1 lbs with VPEs compared to 41,310.4 lbs without VPEs. Incorporating VPEs into the system led to a marginal increase in both the total operating cost and emissions.

4.2.4. Test System IV: Bi-Objective Problem with VPEs and Ramp Rate Constraints

It is already seen that the incorporation of VPEs increases both costs and emissions, whereas the ramp rate constraints lead to an increase in emissions and a slight reduction in costs. Intuitively, incorporating both ramp rate constraints and VPEs will significantly complicate the optimization landscape. Test system IV renders this assumption, and the results are visualized in Figure 5 compared to test system I (without ramp rates and no VPEs). It is demonstrated that the introduction of ramp rates and the integration of VPEs elevate most of the hourly operating costs and emissions.

4.2.5. Test System V: Tri-Objective Problem without VPEs and No Ramp Rates

Practical applications of the MADRL algorithm on test systems I through IV are all used to solve bi-objective problems. The proposed methodology has also been applied to the tri-objective problem of test system V and yielded insightful results presented in Table 9. Test system V deals with the complex interplay of operating costs, CO2, and SO2 emissions, which is ramp rate constrained. The proposed algorithm has achieved a total operating cost of $568,091.2 (a startup cost of $4100.0 and a production cost of $563,991.2). It has also resulted in CO2 and SO2 emissions of 82,261.0 lbs and 161,403.6 lbs, respectively. Compared with existing methods such as [36], the proposed method has demonstrated notable advantages across the three objectives. MADRL has achieved remarkable efficiency for the cost criterion, with a daily cost of $568,091.2. EAD, NSGA-II, and NSGA-III resulted in higher costs of $573,537.7, $572,768.1, and $568,827.9, respectively. For CO2 emissions, the proposed method has also turned out to be competitive with emissions of 82,261.0 lbs, while the emissions from the EAD, NSGA-II, and NSGA-III methods are 83,039.0 lbs, 82,250.1 lbs, and 81,805.6 lbs, respectively. Furthermore, for SO2 emissions, MADRL has outperformed with emissions of 161,403.6 lbs, while the EAD, NSGA-II, and NSGA-III methods render 159,106.7, 157,393.1, and 167,085.4 lbs of emissions, respectively. The results indicate the competitive edge of MADRL over existing methods [36] like EAD, NSGA-II, and NSGA-III in minimizing both environmental emissions and operating costs simultaneously. Its ability to handle varying constraints and optimize multiple objectives positions it as a highly efficient and promising methodology for addressing complex challenges in real-world power systems. The findings demonstrate the robustness and adaptability of the MADRL approach under various complexities.

4.2.6. Test System VI: Ramp Rates Constrained Tri-Objective Problem

Test system VI is a tri-objective problem like test system V, except it incorporates ramp rate constraints. The proposed algorithm renders a total operating cost of $568,186.3, CO2 emissions of 82,373.9 lbs, and SO2 emissions of 161,196.4 lbs. A direct comparative analysis is impossible without existing literature on this specific configuration. However, the results offer valuable insights compared to the tri-objective problem without ramp rate constraints.

4.2.7. Large Scale Test Systems: Test Systems VII–XXVI

Leveraging the insights acquired from the 10-unit test systems I-VI, the scalability of the proposed algorithm has been further explored in larger-scale problems consisting of 20, 40, 60, 80, and 100 units. The larger test systems are magnified versions of the initial four test systems. To ensure consistency, the load demands are adjusted proportionally to accommodate the increased system size. Thus, each larger test system has four distinct configurations, corresponding to test systems I-IV. Table 10 summarizes the total daily operating costs and emissions of all the bi-objective problems. A noticeable rise in cost and emissions can be observed when both ramp rates and VPEs are introduced in each larger-scale test system.

4.3. Training Convergence

A learning curve is a graphical representation that illustrates the performance of an agent(s) over the training episodes. The training converges when the learning curve stabilizes or forms a plateau. Each of the trained models has been examined for convergence. For instance, Figure 6 illustrates the evolution of the normalized values for the tri-objective problem corresponding to test system V. The plot clearly shows that the value of the tri-objective function stabilizes for episodes over 800. It further reveals that the algorithm consistently achieves the minimum tri-objective value for episodes over 900, indicating training convergence. Moreover, the figure depicts correlations among the separate objectives (cost, CO2, and SO2). A clear positive correlation between CO2 emissions and economic costs across the training episodes is visible. Conversely, SO2 emissions are inversely related to costs and CO2 emissions. This underscores the intricate trade-offs between the two types of emissions, where minimizing one type of emission increases the other.

5. Concluding Remarks

This article introduces a novel MADRL algorithm designed for solving the MOPS problems, which integrates economic costs and several types of environmental emissions. The formulation includes bi-objective scenarios as well as tri-objective problems involving cost, CO2, and SO2 emissions. The effectiveness of the algorithm is empirically demonstrated across varying scales and characteristics of power systems. The results revealed that MADRL performs better than established methods such as TLBO, RCGWO, EAD, NSGA-II, and NSGA-III. The algorithm’s successful implementation in 20- to 100-unit systems with varying complexities highlights its scalability in handling more extensive power system scheduling challenges while ensuring efficient and eco-friendly solutions. The proposed algorithm offers flexibility as it is not necessarily restricted to a limited planning horizon and a fixed number of generating units. The simulation environment is also designed to be model agnostic rather than tailored to a specific RL model. This adaptability allows researchers and practitioners to train and explore diverse types of MADRL models for solving power scheduling. Our proposed algorithm’s current focus on thermal generating units marks the initial step toward a more sustainable energy future. The algorithm’s potential to integrate renewable energy sources such as solar, wind, and hydropower opens the door to eco-conscious energy management. Additionally, while the proposed algorithm adeptly handles bi- and tri-objective power scheduling problems, the complexity of modern power systems often presents multifaceted objectives. Therefore, the logical progression for this research lies in further development to accommodate more than three objectives.

Author Contributions

Conceptualization, Y.J.K. and A.S.E.; methodology, Y.J.K.; software, A.S.E.; validation, Y.J.K.; formal analysis, A.S.E.; investigation, A.S.E.; resources, Y.J.K.; data curation, A.S.E.; writing—original draft preparation, A.S.E.; writing—review and editing, Y.J.K.; visualization, A.S.E.; supervision, Y.J.K.; project administration, Y.J.K.; funding acquisition, Y.J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF), funded by the Ministry of Education (2021R1I1A3047456).

Data Availability Statement

The data used can be found in the cited references in Section 4.1.1.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Nomenclature

Indices
T :Number of periods in a scheduling horizon.
n :Number of power-generating units.
m :Number of types of emissions
T = 1,2 , , T :Indices of all periods, t T .
T = 1,2 , , n :Indices of all units, i T .
K = 0,1 , , m :Indices of all objectives, k K
Supply and Demand Profiles
p i m a x , p i m i n :Maximum, minimum capacity of unit i (MW).
p i u p ,   p i d o w n :Maximum ramp up, ramp down unit i (MW).
t i u p ,   t i d o w n :Minimum up-, down-time duration of the unit i (hour).
t i c o l d :Number of cold start periods (hour)
C i h o t , C i c o l d :Hot, cold start cost ($)
p t i :Power output of unit i at period t (MW).
t t i O N ,   t t i O F F :Online, offline duration of unit i at period t (hour).
t t i : Operating (online/offline) duration of unit i at period t (hour).
d t :Demand at period t (MW).
r :Proportion of demand for reserve capacity.
Cost Functions
C :Total operating cost ($) in the entire planning horizon.
C t :Total operating cost ($) at period t .
C t i s u ,   C o n p t i ,   C t i s d :Startup, production, shutdown cost ($) of unit i at period t .
α i c ,   β i c ,   δ i c :   Quadratic, linear, constant cost parameters of unit i .
ρ i c , φ i c :   Valve point cost parameters of the unit i .
Emission Functions
E :Total operating emissions (lbs) in the entire planning horizon.
E t :Total operating emissions (lbs) at period t .
E t i s u ,   E o n p t i ,   E t i s d :Startup, production, shutdown emission (lbs) of unit i at period t .
α i e ,   β i e ,   δ i e :   Quadratic, linear, constant emission parameters of unit i .
ρ i e , φ i e :   Valve point emission parameters of unit i .
α i C O 2 ,   β i C O 2 ,   δ i C O 2 :   Quadratic, linear, constant CO2 parameters of unit i .
α i S O 2 ,   β i S O 2 ,   δ i S O 2 :   Quadratic, linear, constant SO2 parameters of unit i .
Other Notations
E :Expected value.
R :Set of real numbers.
:Indicator function.

References

  1. Huang, Y.; Pardalos, P.; Zheng, Q. Electrical power unit commitment: Deterministic and two-stage stochastic programming models and algorithms. In Springer Briefs in Energy; Springer: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  2. Goyal, S.; Singh, J.; Saraswat, A.; Kanwar, N.; Shrivastava, M.; Mahela, O. Economic Load Dispatch with Emission and Line Constraints using Biogeography Based Optimization Technique. In Proceedings of the 2020 International Conference on Intelligent Engineering and Management (ICIEM), London, UK, 17–19 June 2020. [Google Scholar]
  3. Rex, C.R.E.S.; Marsaline, B.M. State of art in combined economic and emission dispatch. Middle-East J. Sci. Res. 2017, 25, 56–64. [Google Scholar]
  4. Montero, L.; Bello, A.; Reneses, J. Review on the Unit Commitment Problem: Approaches, Techniques, and Resolution Methods. Energies 2022, 15, 1296. [Google Scholar] [CrossRef]
  5. Qin, J.; Yu, N.; Gao, Y. Solving Unit Commitment Problems with Multi-step Deep Reinforcement Learning. In Proceedings of the 2021 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), Virtually, 25–28 October 2021. [Google Scholar]
  6. de Oliveira, L.; da Silva Junior, I.; Abritta, R. Search Space Reduction for the Thermal Unit Commitment Problem through a Relevance Matrix. Energies 2022, 15, 7153. [Google Scholar] [CrossRef]
  7. Bendotti, P.; Fouilhoux, P.; Rottner, C. On the complexity of the unit commitment problem. Ann. Oper. Res. 2019, 274, 119–130. [Google Scholar] [CrossRef]
  8. Roy, P.; Roy, P.; Chakrabarti, A. Modified shuffled frog leaping algorithm with genetic algorithm crossover for solving economic load dispatch problem with valve-point effect. Appl. Soft Comput. 2013, 13, 4244–4252. [Google Scholar] [CrossRef]
  9. Wang, T.; He, X.; Huang, T.; Li, C.; Zhang, W. Collective neurodynamic optimization for economic emission dispatch problem considering valve point effect in microgrid. Neural Netw. 2017, 93, 126–136. [Google Scholar] [CrossRef]
  10. Wang, C.; Shahidehpour, S. Effects of ramp rate Limits on unit commitment and Economic Dispatch. IEEE Trans. Power Syst. 1993, 8, 1341–1350. [Google Scholar] [CrossRef]
  11. Zaoui, S.; Belmadani, A. Solution of combined economic and emission dispatch problems of power systems without penalty. Appl. Artif. Intell. 2022, 36, 1976092. [Google Scholar] [CrossRef]
  12. Jasmin, E.; Ahamed, T.; Remani, T. A function approximation approach to reinforcement learning for solving unit commitment problem with photo voltaic sources. In Proceedings of the 2016 IEEE International Conference on Power Electronics, Drives and Energy Systems, Kerala, India, 14–17 December 2016. [Google Scholar]
  13. Park, H. A Unit Commitment Model Considering Feasibility of Operating Reserves under Stochastic Optimization Framework. Energies 2022, 15, 6221. [Google Scholar] [CrossRef]
  14. Feng, Z.K.; Niu, W.J.; Wang, W.C.; Zhou, J.Z.; Cheng, C.T. A mixed integer linear programming model for unit commitment of thermal plants with peak shaving operation aspect in regional power grid lack of flexible hydropower energy. Energy 2019, 175, 618–629. [Google Scholar] [CrossRef]
  15. Lin, S.; Wu, H.; Liu, J.; Liu, W.; Liu, Y.L.M. A Solution Method for Many-Objective Security-Constrained Unit Commitment Considering Flexibility. Front. Energy Res. 2022, 10, 857520. [Google Scholar] [CrossRef]
  16. Srikanth, K.; Panwar, L.; Panigrahi, B.; Herrera-Viedma, E.; Sangaiah, A.; Wang, G. Meta-heuristic framework: Quantum inspired binary grey wolf optimizer for unit commitment problem. Comput. Electr. Eng. 2018, 70, 243–260. [Google Scholar] [CrossRef]
  17. Yang, Z.; Li, K.; Niu, Q.; Xue, Y. A novel parallel-series hybrid meta-heuristic method for solving a hybrid unit commitment problem. Knowl. Based Syst. 2017, 134, 13–30. [Google Scholar] [CrossRef]
  18. Kigsirisin, S.; Miyauchi, H. Short-Term Operational Scheduling of Unit Commitment Using Binary Alternative Moth-Flame Optimization. IEEE Access 2021, 9, 12267–12281. [Google Scholar] [CrossRef]
  19. Trivedi, A.; Srinivasan, D.; Biswas, S.; Reindl, T. Hybridizing genetic algorithm with differential evolution for solving the unit commitment scheduling problem. Swarm Evol. Comput. 2015, 23, 50–64. [Google Scholar] [CrossRef]
  20. Zhu, X.; Zhao, S.; Yang, Z.; Zhang, N.; Xu, X. A parallel meta-heuristic method for solving large scale unit commitment considering the integration of new energy sectors. Energy 2022, 238, 121829. [Google Scholar] [CrossRef]
  21. Basu, M. Economic environmental dispatch using multi-objective differential evolution. Appl. Soft Comput. 2011, 11, 2845–2853. [Google Scholar] [CrossRef]
  22. Ponciroli, R.; Stauff, N.; Ramsey, J.; Ganda, F.; Vilim, R. An improved genetic algorithm approach to the unit commitment/economic dispatch problem. IEEE Trans. Power Syst. 2020, 35, 4005–4013. [Google Scholar] [CrossRef]
  23. Balasubramanian, K.; Santhi, R. Best Compromised Schedule for Multi-Objective Unit Commitment Problems. Indian J. Sci. Technol. 2016, 9, 2. [Google Scholar] [CrossRef]
  24. Roy, P.; Sarkar, R. Solution of unit commitment problem using quasi-oppositional teaching learning based algorithm. Electr. Power Energy Syst. 2014, 60, 96–106. [Google Scholar] [CrossRef]
  25. Datta, D. Unit commitment problem with ramp rate constraint using a binary-real-coded genetic algorithm. Appl. Soft Comput. 2013, 13, 3873–3883. [Google Scholar] [CrossRef]
  26. Reddy, S.; Kumar, R.; Panigrahi, B. Binary Bat Search Algorithm for Unit Commitment Problem in Power system. In Proceedings of the IEEE International WIE Conference on Electrical and Computer Engineering (WIECON-ECE), Dehradun, India, 14–16 December 2017. [Google Scholar]
  27. Khunkitti, S.; Watson, N.; Chatthaworn, R.; Premrudeepreechacharn, S.; Siritaratiwat, A. An Improved DA-PSO Optimization Approach for Unit Commitment Problem. Energies 2019, 12, 2335. [Google Scholar] [CrossRef]
  28. Panwar, L.; Reddy, S.; Verma, A.; Panigrahi, B.; Kumar, R. Binary grey wolf optimizer for large scale unit commitment problem. Swarm Evol. Comput. 2018, 38, 251–266. [Google Scholar] [CrossRef]
  29. Li, F.; Qin, J.; Zheng, W. Distributed Q-learning-based online optimization algorithm for unit commitment and dispatch in smart grid. IEEE Trans. Cybern. 2020, 50, 4146–4156. [Google Scholar] [CrossRef]
  30. Reddy, S.P.L.; Panigrahi, B.; Kumar, R. Solution to unit commitment in power system operation planning using binary coded modified moth flame optimization algorithm (BMMFOA): A flame selection based computational technique. J. Comput. Sci. 2018, 25, 298–317. [Google Scholar]
  31. Nassef, A.A.M.; Maghrabie, H.; Baroutaji, A. Review of Metaheuristic Optimization Algorithms for Power. Sustainability 2023, 15, 9434. [Google Scholar] [CrossRef]
  32. Kumar, V.; Kumar, D. Binary whale optimization algorithm and its application to unit commitment problem. Neural Comput. Appl. 2020, 32, 2095–2123. [Google Scholar] [CrossRef]
  33. Zhai, Y.; Liao, X.; Mu, N.; Le, J. A two-layer algorithm based on PSO for solving unit commitment problem. Soft Comput. 2020, 24, 9161–9178. [Google Scholar] [CrossRef]
  34. Rameshkumar, J.; Ganesan, S.; Abirami, M.; Subramanian, S. Cost, emission and reserve pondered predispatch of thermal power generating units coordinated with real coded grey wolf optimization. IET Gener. Transm. Distrib. 2016, 10, 972–985. [Google Scholar] [CrossRef]
  35. Bora, T.C.; Mariani, V.C.; dos Santos Coelho, L. Multiobjective optimization of the environmental-economic dispatch with reinforcement learning based on non-dominated sorting genetic algorithm. Appl. Therm. Eng. 2019, 146, 688–700. [Google Scholar] [CrossRef]
  36. Yang, D.; Zhou, X.; Yang, Z.; Guo, Y.; Niu, Q. Low Carbon Multi-Objective Unit Commitment Integrating Renewable Generations. IEEE Access 2020, 8, 207768–207778. [Google Scholar] [CrossRef]
  37. Trivedi, A.; Srinivasan, D.; Pal, K.; Saha, C.; Reindl, T. Enhanced Multiobjective Evolutionary Algorithm based on Decomposition for Solving the Unit Commitment Problem. IEEE Trans. Ind. Inform. 2009, 11, 1346–1357. [Google Scholar] [CrossRef]
  38. Wang, B.; Wang, S.; Zhou, X.; Watada, J. Two-Stage Multi-Objective Unit Commitment Optimization Under Hybrid Uncertainties. IEEE Trans. Power Syst. 2015, 31, 2266–2277. [Google Scholar] [CrossRef]
  39. Wang, B.; Wang, S.; Zhou, X.; Watada, J. Multi-objective unit commitment with wind penetration and emission concerns under stochastic and fuzzy uncertainties. Energy 2016, 111, 18–31. [Google Scholar] [CrossRef]
  40. de Mars, P.; O’Sullivan, A. Applying reinforcement learning and tree search to the unit commitment problem. Appl. Energy 2021, 302, 117519. [Google Scholar] [CrossRef]
  41. Rajasomashekar, S.; Aravindhababu, P. Biogeography based optimization technique for best compromise solution of economic emission dispatch. Swarm Evol. Comput. 2012, 7, 47–57. [Google Scholar] [CrossRef]
  42. Ebrie, A.; Paik, C.; Chung, Y.; Kim, Y. Environment-Friendly Power Scheduling Based on Deep Contextual Reinforcement Learning. Energies 2023, 16, 5920. [Google Scholar] [CrossRef]
  43. Jasmin, E.; Ahamed, T.; Jagathy, R. Reinforcement learning solution for unit commitment problem through pursuit method. In Proceedings of the 2009 International Conference on Advances in Computing, Control, and Telecommunication Technologies, Kerala, India, 28–29 December 2009. [Google Scholar]
  44. Rajua, L.; Milton, R.; Suresha, S.; Sankara, S. Reinforcement Learning in Adaptive Control of Power System Generation. Procedia Comput. Sci. 2015, 46, 202–209. [Google Scholar] [CrossRef]
  45. Navin, N.; Sharma, R. A fuzzy reinforcement learning approach to thermal unit commitment problem. Neural Comput. Appl. 2019, 31, 737–750. [Google Scholar] [CrossRef]
  46. de Mars, P.; O’Sullivan, A. Reinforcement learning and A* search for the unit commitment problem. Energy AI 2022, 9, 100179. [Google Scholar] [CrossRef]
  47. Dalal, G.; Mannor, S. Reinforcement learning for the unit commitment problem. In Proceedings of the 2015 IEEE Eindhoven PowerTech, Eindhoven, The Netherlands, 29 June–2 July 2015. [Google Scholar]
  48. Ebrie, A.; Kim, Y. pymops: A multi-agent simulation-based optimization package for power scheduling. Softw. Impacts 2024, 19, 1006160. [Google Scholar] [CrossRef]
  49. Ongsakul, W.; Petcharaks, N. Unit commitment by enhanced adaptive Lagrangian relaxation. IEEE Trans. Power Syst. 2004, 19, 620–628. [Google Scholar] [CrossRef]
  50. Walters, D.; Sheble, G. Genetic algorithm solution of economic dispatch with valve point loading. IEEE Trans. Power Syst. 1993, PAS 97, 1325–1332. [Google Scholar] [CrossRef]
  51. Guesmi, T.; Farah, A.; Marouani, I.; Alshammari, B.; Abdallah, H. Chaotic sine–cosine algorithm for chance-constrained economic emission dispatch problem including wind energy. IET Renew. Power Gener. 2020, 14, 1801–1808. [Google Scholar] [CrossRef]
  52. Li, L.-L.; Lou, J.-L.; Tseng, M.-L.; Lim, M.; Tan, R. A hybrid dynamic economic environmental dispatch model for balancing operating costs and pollutant emissions in renewable energy: A novel improved mayfly algorithm. Expert Syst. Appl. 2022, 203, 117411. [Google Scholar] [CrossRef]
  53. Dey, S.; Dash, D.; Basu, M. Application of NSGA-II for environmental constraint economic dispatch of thermal-wind-solar power system. Renew. Energy Focus 2022, 43, 239–245. [Google Scholar]
  54. Li, Y.; Pedroni, N.; Zio, E. A memetic evolutionary multiobjective optimization method for environmental power unit commitment. IEEE Trans. Power Syst. 2013, 28, 2660–2669. [Google Scholar] [CrossRef]
  55. Sutton, R.; Barto, A. Reinforcement Learning, An Introduction; The MIT Press: London, UK, 2018. [Google Scholar]
  56. Virtanen, P.; Gommers, R.; Oliphant, T.S. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [PubMed]
  57. Adam, S.; Busoniu, L.; Babuska, R. Experience Replay for Real-Time Reinforcement Learning Control. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2012, 42, 201–212. [Google Scholar] [CrossRef]
  58. Attaviriyanupap, P.; Kita, H.; Tanaka, E.; Hasegawa, J. A Hybrid EP and SQP for Dynamic Economic Dispatch with Nonsmooth Fuel Cost Function. IEEE Trans. Power Syst. 2002, 17, 2. [Google Scholar] [CrossRef]
  59. Truby, J. Thermal Power Plant Economics and Variable Renewable Energies: A Model-Based Case Study for Germany; International Energy Agency: Paris, France, 2014. [Google Scholar]
Figure 1. Flow chart of the MOPS simulation dynamics as MDPs.
Figure 1. Flow chart of the MOPS simulation dynamics as MDPs.
Systems 12 00106 g001
Figure 2. Interaction between the multi-agent simulation environment and the MADQN.
Figure 2. Interaction between the multi-agent simulation environment and the MADQN.
Systems 12 00106 g002
Figure 3. Plot of hourly costs and emissions between RCGWO and MADRL for test system I.
Figure 3. Plot of hourly costs and emissions between RCGWO and MADRL for test system I.
Systems 12 00106 g003
Figure 4. Plot of hourly costs and emissions between RCGWO and MADRL for test system II.
Figure 4. Plot of hourly costs and emissions between RCGWO and MADRL for test system II.
Systems 12 00106 g004
Figure 5. Plot of hourly costs and emissions between test systems I and IV.
Figure 5. Plot of hourly costs and emissions between test systems I and IV.
Systems 12 00106 g005
Figure 6. Mean evolution of normalized values for test system V across 10,000 training episodes.
Figure 6. Mean evolution of normalized values for test system V across 10,000 training episodes.
Systems 12 00106 g006
Table 2. Estimates of cost-to-emission conversion parameters of generating units.
Table 2. Estimates of cost-to-emission conversion parameters of generating units.
η U1U2U3U4U5U6U7U8U9U10
Without VPEs C - to - E 1.121.121.181.721.321.601.420.630.640.63
C - to - E C O 2 1.261.311.971.471.361.802.002.582.353.42
C - to - E S O 2 4.644.984.003.745.233.517.162.642.632.67
With VPEs C - to - E 1.091.111.151.641.291.531.420.630.620.60
Table 3. Top ten weights with the largest cardinality values for the bi-objective problems.
Table 3. Top ten weights with the largest cardinality values for the bi-objective problems.
Test
System
12345678910
I ω 0.350.360.370.380.390.400.410.420.430.44
μ C 0.8780070.8904430.9016380.9125860.9210310.9286620.9356480.9425700.9492120.955643
μ E 0.2073310.2730580.2527670.2320820.2154220.1997290.1847790.1693050.1538490.138279
μ C , E 0.0098820.0105930.0105100.0104220.0103470.0102740.0102010.0101230.0100430.009960
II ω 0.370.380.390.400.410.420.430.440.450.46
μ C 0.8982680.9097010.9178720.9253320.9324280.9390340.9456120.9519020.9575160.953013
μ E 0.1613890.2364370.2206930.2056210.1906790.1761480.1611540.1461860.1323350.135774
μ C , E 0.0098760.0106820.0106110.010540.0104670.0103930.0103150.0102340.0101570.010147
III ω 0.300.360.370.400.450.480.620.910.920.95
μ C 0.0000010.0000010.0409600.0756240.0832360.0989140.2522960.2072960.3063730.260789
μ E 0.3114850.3010390.2732150.2373540.2202390.2185290.0584570.1096960.0117300.043777
μ C , E 0.0098210.0094920.0099060.0098680.0095690.0100090.0097980.0099950.0100300.009603
IV ω 0.420.540.680.690.700.730.870.960.970.98
μ C 0.1444620.1101420.2133090.2168200.1863390.2080770.1967140.2226550.2250960.164261
μ E 0.0876510.1159800.0135100.0154740.0000010.0000010.0000010.0067460.0000010.000001
μ C , E 0.0068310.0066550.0066760.0068370.0054840.0061240.0057900.0067520.0066250.004834
Table 4. Top ten weights with the largest cardinality values for the tri-objective problems.
Table 4. Top ten weights with the largest cardinality values for the tri-objective problems.
Test
System
12345678910
V ω c o s t 0.070.130.150.170.310.430.460.470.560.61
ω C O 2 0.670.600.570.520.380.210.220.170.060.06
ω S O 2 0.260.270.280.310.310.360.320.360.380.33
μ C 0.8775440.8747370.8768520.8670230.8762700.8667990.8825100.8699050.8633710.884918
μ E C O 2 0.5158610.5183070.5258210.5138060.5499590.5603090.5831000.5722890.5796970.612661
μ E S O 2 0.6171230.6252200.6208350.6431780.6243590.6446000.6104280.6379670.6500020.603679
μ C , E 0.0051760.0051750.0052110.0051300.0052980.0053020.0054450.0053580.0053610.005563
VI ω c o s t 0.030.210.240.410.430.440.450.490.520.52
ω C O 2 0.760.520.470.280.210.230.190.150.120.12
ω S O 2 0.210.260.290.310.360.330.350.360.360.36
μ C 0.8905290.8851260.8788970.8823600.8668620.8781710.8712100.8705200.8720000.871944
μ E C O 2 0.5309250.5486830.5427290.5751600.5617620.5736720.5698080.5761170.5821080.583029
μ E S O 2 0.5806500.6016610.6174810.6107050.6444210.6204140.6353850.6365380.6332000.633290
μ C , E 0.0052960.0053420.0052970.0054310.0053230.0054100.0053690.0053900.0054180.005421
Table 5. Parameters and optimal hyperparameters of the MADQN models.
Table 5. Parameters and optimal hyperparameters of the MADQN models.
Test
System
ObjectivesUnitsNodes -Decay
Rate
Replay
Memory
Training
Episodes
InputHiddenOutput
IBi-objective101264200.999648000
IIBi-objective101264200.999648000
IIIBi-objective101264200.999648000
IVBi-objective101264200.999648000
VTri-objective101264200.99916410,000
VITri-objective101264200.99916410,000
VII–XBi-objective202264400.999648000
XI–XIVBi-objective4042128800.999312812,000
XV–XVIIIBi-objective60621281200.999312812,000
XIX–XXIIBi-objective80822561600.999425615,000
XXIII–XXVIBi-objective1001022562000.999425615,000
Table 6. Optimal commitments, loads, and reserve percentages using MADRL for test system I.
Table 6. Optimal commitments, loads, and reserve percentages using MADRL for test system I.
t
(Hour)
CommitmentsOptimal Loads (MW) r %
p t 1 p t 2 p t 3 p t 4 p t 5 p t 6 p t 7 p t 8 p t 9 p t , 10
11100000000375.8324.20000000030.0
21100000000400.6349.40000000021.3
31100100000405.5354.30090.20000026.1
41100100000442.5391.700115.90000012.8
51101100000412.9361.80130.095.40000020.2
61111100000401.8350.5130.0130.087.70000021.1
71111100000420.3369.2130.0130.0100.50000015.8
81111100000438.8387.9130.0130.0113.30000011.0
91111111000455.0409.4130.0130.0128.021.725.900015.2
101111111100455.0441.4130.0130.0149.937.728.527.50010.9
111111111110455.0454.1130.0130.0158.744.129.529.618.9010.8
121111111111455.0455.0130.0130.0162.058.431.834.523.919.310.8
131111111100455.0441.4130.0130.0149.937.728.527.50010.9
141111111000455.0409.4130.0130.0128.021.725.900015.2
151111100000438.8387.9130.0130.0113.30000011.0
161111100000383.3331.8130.0130.074.80000026.9
171111100000364.8313.1130.0130.062.00000033.2
181111100000401.8350.5130.0130.087.70000021.1
191111100000438.8387.9130.0130.0113.30000011.0
201111111100455.0441.4130.0130.0149.937.728.527.50010.9
211111111000455.0409.3130.0130.0128.021.725.900015.2
221100111000455.0435.800146.135.028.100012.5
231100100000424.0373.000103.10000019.1
241100000000425.5374.50000000013.7
Table 7. Comparison of costs and emissions between RCGWO and MADRL for test system I.
Table 7. Comparison of costs and emissions between RCGWO and MADRL for test system I.
t
(Hour)
RCGWO [34]MADRL
C t s u ($) C t o n ($) C t ($) E t (lbs) C t s u ($) C t o n ($) C t ($) E t (lbs)
1013,683.113,683.1956.4013,750.313,750.3861.1
2014,554.514,554.51055.0014,601.214,601.21002.2
356016,892.117,452.11077.490017,027.517,927.51108.7
455019,261.519,811.51249.8018,821.218,821.21373.8
5020,132.520,132.51343.856020,246.320,806.31243.5
690022,387.123,287.115,52.7110022,601.023,701.01394.3
7023,262.023,262.01704.2023,496.623,496.61521.4
8024,150.324,150.31863.1024,394.024,394.01656.3
986027,251.128,111.12191.386027,399.128,259.12085.1
106030,057.630,117.62599.26030,226.230,286.22490.5
116031,916.131,976.12945.26032,045.632,105.62843.0
126033,890.233,950.23229.46033,995.634,055.63146.4
13030,057.630,057.62599.2030,226.230,226.22490.5
14027,251.127,251.12191.3027,399.127,399.12085.1
15024,150.324,150.31863.1024,394.024,394.01656.3
16021,513.721,513.71424.2021,707.321,707.31274.9
17020,641.820,641.81318.6020,815.420,815.41163.3
18022,387.122,387.11552.7022,601.022,601.01394.3
19024,150.324,150.31863.1024,394.024,394.01656.3
2049030,057.630,547.62599.249030,226.230,716.22490.5
21027,251.127,251.12191.3027,399.127,399.12085.1
22023,593.023,593.01719.6022,847.122,847.11895.6
23019,480.819,480.81348.9017,923.517,923.51237.4
24017,142.817,142.81321.1015,453.115,453.11154.8
Total3540565,115.3568,655.343,759.74090.0563,990.6568,080.641,310.4
Table 8. Comparison of costs and emissions between RCGWO and MADRL for test system II.
Table 8. Comparison of costs and emissions between RCGWO and MADRL for test system II.
t
(Hour)
RCGWO [34]MADRL
C t s u ($) C t o n ($) C t ($) E t (lbs) C t s u ($) C t o n ($) C t ($) E t (lbs)
1013,683.013,683.0956.4013,748.313,748.3862.1
2014,554.414,554.41055.0014,599.314,599.31003.2
356016,892.017,452.01077.390016,998.917,898.91125.3
455019,261.419,811.41249.8018,789.818,789.81392.3
5020,132.420,132.41343.856020,217.120,777.11260.5
690022,387.223,287.21552.7110022,572.723,672.71410.7
7023,262.123,262.11704.3023,466.823,466.81538.7
8024,150.124,150.11863.1024,362.924,362.91674.5
986027,251.328,111.32191.386027,365.028,225.02103.9
106030,057.830,117.82599.26030,128.430,188.42545.2
116031,916.331,976.32945.26032,008.332,068.32864.0
126033,890.433,950.43229.46033,972.934,032.93159.7
13030,057.830,057.82599.2030,128.530,128.52545.2
14027,251.327,251.32191.3027,365.027,365.02103.9
15024,150.124,150.11863.1024,362.924,362.91674.5
16021,513.821,513.81424.2021,680.321,680.31290.4
17020,642.020,642.01318.6020,789.720,789.71177.9
18022,387.222,387.21552.7022,572.722,572.71410.7
19024,150.124,150.11863.1024,362.924,362.91674.5
2049030,124.930,614.92573.149030,128.530,618.52545.2
21027,251.327,251.32191.3027,365.027,365.02103.9
22023,593.123,593.11719.6022,776.222,776.21933.8
23019,481.019,481.01349.0017,893.517,893.51254.9
24017,142.917,142.91321.1015,451.215,451.21155.8
Total3540565,183.9568,723.943,733.64090563,106.8567,196.841,810.8
Table 9. Optimal loads, reserve percentages, costs, and CO2 and SO2 emissions for test system V.
Table 9. Optimal loads, reserve percentages, costs, and CO2 and SO2 emissions for test system V.
t
(Hour)
Optimal Loads (MW) r % C t ($) E t C O 2 (lbs) E t S O 2 (lbs)
p t 1 p t 2 p t 3 p t 4 p t 5 p t 6 p t 7 p t 8 p t 9 p t , 10
1357.9342.10000000030.013,766.81891.33466.5
2382.1367.90000000021.314,618.22001.84087.5
3391.8378.30079.90000026.117,910.02380.64703.5
4434.2423.40092.40000012.818,757.82606.36063.5
5400.3387.3130.0082.40000020.220,801.12896.15460.1
6391.3377.8130.0121.279.70000021.123,711.73305.75884.6
7410.0397.7130.0127.185.20000015.823,462.33421.56532.2
8430.0418.9130.0130.091.10000011.024,335.33536.17223.2
9442.3432.1130.0130.094.845.925.000015.228,285.24045.87944.8
10455.0455.0130.0130.0117.259.727.725.40010.930,273.54502.89026.7
11455.0455.0130.0130.0127.365.834.029.923.1010.832,210.04854.79420.4
12455.0455.0130.0130.0134.970.538.933.226.525.910.834,188.25230.39752.3
13455.0455.0130.0130.0117.259.727.725.40010.930,213.54502.89026.7
14442.3432.1130.0130.094.845.925.000015.227,425.24045.87944.8
15430.0418.9130.0130.091.10000011.024,335.33536.17223.2
16374.1359.4126.1115.774.60000026.921,723.13188.95289.7
17357.4341.6120.9110.469.70000033.220,855.73072.24734.0
18391.3377.8130.0121.279.70000021.122,591.73305.75884.6
19430.0418.9130.0130.091.10000011.024,335.33536.17223.2
20455.0455.0130.0130.0117.259.727.725.40010.930,703.54502.89026.7
21442.3432.1130.0130.094.845.925.000015.227,425.24045.87944.8
22455.0455.000109.955.125.000012.522,808.63245.97418.8
23413.0400.90086.10000019.117,883.42493.25360.9
24406.3393.70000000013.715,470.62112.74760.9
Total 568,091.282,261.0161,403.6
Table 10. Results of bi-objective problems with varying cost and emissions functions using MADRL.
Table 10. Results of bi-objective problems with varying cost and emissions functions using MADRL.
Test
Systems
UnitsObjectiveTotal Cost and Emissions
No Ramp Rates, No VPEsRamp Rates, No VPEsNo Ramp Rates, VPEsRamp Rates, VPEs
VII–X20Cost ($)1,136,746.41,135,081.01,179,295.21,184,760.6
Emission (lbs)84,786.085,731.093,047.389,305.9
XI–XIV40Cost ($)2,283,893.82,280,510.12,371,344.62,372,688.8
Emission (lbs)173,154.7175,069.2189,596.4182,460.3
XV–XVIII60Cost ($)3,411,158.63,406,502.83,527,471.53,554,833.5
Emission (lbs)257,715.8260,352.8288,407.1276,125.6
XIX–XXII80Cost ($)4,558,451.24,551,960.64,710,657.64,734,809.3
Emission (lbs)346,550.7350,220.1385,657.1366,742.4
XXIII–XXVI100Cost ($)5,704,920.75,696,950.65,901,065.75,901,065.7
Emission (lbs)434,073.9438,576.0481,106.2481,106.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ebrie, A.S.; Kim, Y.J. Reinforcement Learning-Based Multi-Objective Optimization for Generation Scheduling in Power Systems. Systems 2024, 12, 106. https://doi.org/10.3390/systems12030106

AMA Style

Ebrie AS, Kim YJ. Reinforcement Learning-Based Multi-Objective Optimization for Generation Scheduling in Power Systems. Systems. 2024; 12(3):106. https://doi.org/10.3390/systems12030106

Chicago/Turabian Style

Ebrie, Awol Seid, and Young Jin Kim. 2024. "Reinforcement Learning-Based Multi-Objective Optimization for Generation Scheduling in Power Systems" Systems 12, no. 3: 106. https://doi.org/10.3390/systems12030106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop