Next Article in Journal
Development of a New Sensor Module for an Enhanced Fuel Flexible Operation of Biomass Boilers
Next Article in Special Issue
Influence of Synchronous Condensers on Operation Characteristics of Double-Infeed LCC-HVDCs
Previous Article in Journal
Evaluation of Rate of Adhesion of Lactobacillus namurensis Strain GYP-74 to Porous Fine Ceramics
Previous Article in Special Issue
The Impact of Attacks in LEM and Prevention Measures Based on Forecasting and Trust Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Residential Demand Response Strategy Based on Deep Deterministic Policy Gradient

1
School of Control and Computer Engineering, North China Electric Power University, No. 2 Beinong Road, Changping District, Beijing 102206, China
2
China Electric Power Research Institute, No. 15, Qinghe Xiaoying Road, Beijing 100192, China
*
Author to whom correspondence should be addressed.
Processes 2021, 9(4), 660; https://doi.org/10.3390/pr9040660
Submission received: 15 March 2021 / Revised: 6 April 2021 / Accepted: 6 April 2021 / Published: 9 April 2021

Abstract

:
With the continuous improvement of the power system and the deepening of electricity market reform, the trend of users’ active participation in power distribution is more and more significant. Demand response has become the promising focus of smart grid research. Providing reasonable incentive strategies for power grid companies and demand response strategies for customers plays a crucial role in maximizing the benefits of different participants. To meet different expectations of multiple agents in the same environment, deep reinforcement learning was adopted. The generative model of residential demand response strategy under different incentive policies can be trained iteratively through real-time interactions with the environmental conditions. In this paper, a novel optimization model of residential demand response strategy, based on a deep deterministic policy gradient (DDPG) algorithm, was proposed. The proposed work was validated with the actual electricity consumption data of a certain area in China. The results showed that the DDPG model could optimize residential demand response strategy under certain incentive policies. In addition, the overall goal of peak load-cutting and valley filling can be achieved, which reflects promising prospects of the electricity market.

1. Introduction

Opening the electricity market is the main intent of the new electric power system reform in China. According to the statistics, there are more than 3000 electricity retailers in China. As the number of power retailers increases, depending on flexible price/incentive mechanism users can stimulate the demand response potential of electricity sales companies and the wholesale smart grid to improve the efficiency. Moreover, the proportion of residential electricity consumption in China is constantly growing in recent years. For example, the peak load of residential electricity consumption in Jiangsu province was greater than 1/3 span in the summer of 2020. Besides, the rapid popularization of smart home products has greatly enhanced the potential of residents’ demand response. Therefore, the optimization of large-scale smart residential users’ electricity usage strategy in a complex market environment has important theoretical and practical values.
In order to effectively promote clean energy consumption, the Electric Internet of Things (EIOT) technology needs to be introduced. However, the demand for adjustable capacity has gradually expanded, due to the unbalanced outputs of clean energies connected to the power grid. One of the most effective strategies to solve the uncertain integration of large-scale renewable energy in the grid is aggregating controllable loads on the user side in real-time [1]. With the introduction of the market concept into the power system, attention paid to the demand side and the interactive demand of the power system have been raised to a new level, and participants in the market showed a significant trend of diversification [2]. At the same time, more and more emerging market players and load aggregators appear in front of market participants [3]. The residential load has the characteristics of small units, large amounts, and scattered locations. The main difficulties for electricity sales to optimize power consumption operations are how to aggregate and manage more residents to participate in the grid interactions in a more comfortable way. In general, this includes the optimal decision of electricity purchase and sale and feasible load scheduling plans for users.
To optimize the operation of electricity sales companies to maximize profits, a well-designed interactive model is indispensably considering the load profiles of residents and the decision-making of electricity trading based on the demand response in the electricity market. At present, some exploratory studies have been carried out. Reference [4] applied the utility function to build the residents’ demand response model, and electricity sales companies motivate users to adjust their electricity consumption patterns by implementing load-cutting compensation. Reference [5] integrated residential air conditioning load into the direct control load for dispatching, and an optimized operating model based on the combination of the distributed power supply and the distribution network constraints is established. Reference [6] designed a response framework for air conditioning load, which allocated incentive costs based on user engagement to reduce the impact on user comfort. Reference [7] provided a novel strategy that users could adjust the load operations to realize the demand response on multi-smart homes with regional real-time electricity price predicting. In Reference [8], load aggregators divided residential loads into three categories according to their operation characteristics. Each category of the load is scheduled by grouping aggregations to maximize its benefits in the day-before market. In reference [9], constraints of resident load were considered. The authors established a multi-user intelligent electricity scheduling model and presented a particle swarm optimization algorithm. Reference [10] adopted quadratic utility function modeling for residential users and built a multi-user demand response model for single-period multi-seller electricity companies by applying a two-layer master-slave game. Reference [11] proposed a distributed control model based on the game theory framework to solve the problem of demand response under many flexible loads in the market environment. Reference [12] proposed a distributed optimization framework for intelligent electricity consumption in residential areas. In the previous references, researchers basically tried to obtain more advanced models to describe the real competitive environment of the instant electricity trading market, which equates with a multi-objective optimization task. However, because of the great difficulties in centralized optimization, the aforementioned studies that depended on grouping aggregation and utility function for describing the residential characteristics, cannot be applied in appliance-level objectives. Some other studies focused on residential cases with lightweight models, but it could not be fully expanded to proper scale communities. References [10,11,12] solved the contradiction between the accuracy and scale of the residential demand response model based on the game theory using the distributed optimization characteristics. Yet, it needs to prove the existence of a pure strategic Nash equilibrium solution, which limits the possible applications of game theory in residential demand response.
In summary, giving back to customers involved in grid interactions by reducing their power consumption costs might have great significance in attracting large-scale customers to participate in demand response, allowing power companies to obtain large-scale adjustable loads. Therefore, making a reasonable incentive electricity price and achieving a balance between expanding the load scale and reducing the incentive cost are the key issues to be solved. A strategy optimized model based on deep deterministic policy gradient (DDPG) algorithm is proposed to solve the retail-price making problem for the service provider, and finally verified the effectiveness of the model by using the real power consumption data. This paper is organized as follows. Section 2 gives a brief review of residential demand response and deep reinforcement learning. Section 3 gives the residential demand response model and Section 4 shows the strategy optimized by DDPG and simulations are conducted. Section 5 analyzes the results, and conclusions are given in Section 6.

2. Related Work

2.1. Residential Demand Response

With the continuous increase in new forms of energy consumption, such as electric vehicles and lithium battery storage, it is possible to use incentive measures to guide users to actively participate in load regulation, such as peak shaving and valley filling. The research aim of residential demand response strategy optimization is to reasonably determine the power consumption involved in demand response based on the market demand and the operator’s incentive policy in order to achieve the maximum benefit.
Reference [13] analyzed the variations of load behaviors on the residential side in detail under the incentives of different electricity prices. The authors discussed how the injected power of the distribution network is regulated by demand response resources after adding highly volatile distributed generation into the distribution network and studied the contribution of demand response in smoothing the fluctuation of distributed generation and reducing the load peak. Reference [14] studied the scheme of reducing household electricity consumption and promoting the two-way interactions between the user side and the grid side. Reference [15] paid more attention to automatic demand response, analyzed the obstacles existing in the implementation of automatic demand response in China, and figured out the potential and development direction of China’s automatic demand response. Reference [16] proposed the design principles of the demand response market and the constituent elements of the demand response project. In addition, it classified the types of demand response markets and demand response projects. References [17,18,19] presented a demand response strategy based on the dynamic price in the smart grid environment. Reference [20] proposed the identification model of user interaction degree and solved the problem of processing the inflection point of a piecewise linear response curve through repeated correction of historical data. Tan, et al. [21] gave an optimization model for the design of time-of-use (TOU) price on the generation side considering the real-time grid electricity in different peak valley peacetime periods. Reference [22] designed an optimal decision-making model of TOU price considering the satisfaction of consumers’ electricity consumption behavior. Reference [23] proposed a bilateral price linkage electricity price model based on the field data of municipal power companies and obtained the optimal TOU price for peak load shifting and valley filling through simulation analysis. Wei, et al. [24] studied the peak-to-valley electricity price of high load energy users, took the wind power consumption capacity into account, established a corresponding peak-valley electricity consumption model with continuous division of peak-to-valley periods, and finally established the peak-to-valley electricity price decision-making model with the goal of increasing the wind power consumption capacity of the system during the valley period. Through decoupling time division and electricity price optimization, a two-step optimization method was designed to build the comprehensive optimization model of peak-to-valley period and electricity price. Reference [25] presented the optimal operation scheme for household user side micro-grid considering the TOU price and demand response. Reference [26] devised the optimization model for large consumers’ direct wind power consumption under TOU price. Reference [27] proposed the multi-objective optimal scheduling of electric vehicle charging and discharging based on TOU price. Fan et al. [28] provided a price optimization strategy for the access time of photovoltaic power generation in the smart grid. Overall, the studies of demand response strategy usually rely on the mathematical models which need more work on the models of participants with this type of formula. It is necessary to look at data-driven methods by using artificial intelligence technology.

2.2. Deep Reinforcement Learning

Reinforcement learning, supervised learning and unsupervised learning constitute the core machine learning section of artificial intelligence technology. Supervised learning is learning from a marked training set. The characteristics of each sample in the training set can be regarded as a description of the scene, and its label can be regarded as the correct action that should be performed. The purpose of unsupervised learning can be defined as finding the hidden rules from an amount of unlabeled samples.
Reinforcement learning runs in a specific situation with the process of trying and choosing which kind of action can get the maximum rewards. The difference between reinforcement learning and other machine learning algorithms is that there is no supervisor in the learning process, where the interactions between actions and related feedbacks occur in this process. Therefore, decision-making behavior will affect a series of states and rewards significantly.
Deep reinforcement learning combines deep learning skills with reinforcement learning. Deep learning is a neural network with hidden multi-layer. Its basic theory is to enhance low-level features through multi-layer network structure and nonlinear transformation and to discover the distinguished representation of the target by acquiring abstract and more discriminative high-level representation. In recent years, deep learning skills, such as the most popular supervised learning, have been successfully applied in smart grids through load forecasting [29], price peak prediction [30], optimization of battery storage [31], etc. Reinforcement learning comes from the behaviorism theory in psychology which produces the habitual behavior that can obtain the maximum benefit under the guide of the reward or punishment scheme given by the environment. Deep reinforcement learning transforms the traditional value iteration into the training of the neural network, breaks the limitation of state dimension and discrete action in the classical reinforcement learning, and uses a deep neural network with a strong fitting ability to establish strategy function [32,33]. Deep reinforcement learning technology has been widely used and plays an important role in games, robot control, dialogue system, autonomous driving, and other fields [34,35,36,37,38,39,40].
Deep reinforcement learning algorithm is described and classified in reference [41], which can be divided into two basic methods based on value function and policy gradient. Typical classifications of deep reinforcement learning algorithms are shown in Table 1, among which deep Q-learning (DQN), advantage actor-critic (A3C), DDPG, and Distributed proximal policy optimization (DPPO) are the most representative deep reinforcement learning models at present.
The DDPG algorithm is a combination of the actor–critic framework and the DQN algorithm. The experience playback mechanism and a separate target network are used to reduce the correlation between the data set and increase the stability and robustness of the algorithm. In addition, considering the problem that the continuous action space is easy to fall into the local optimum, the design of the "noise" item is added, and the exploration efficiency is improved by strengthening the random behaviors in the action strategy. Both the actor and critic dual networks in the DDPG algorithm adopt the target network and evaluate the network design. During the training process, in order to increase the stability of the algorithm learning process, only the parameters of the estimated network need to be updated, and the training steps are separated and copied to the target network.
A systematic review of artificial intelligence and machine learning methods used in DR is reported in [42]. Reinforcement learning [43] especially deep reinforcement learning [44], such as advantage actor-critic (A3C) and deep Q-learning (DQN) [45] are promising approaches for strategy optimization of residential demand response. The DDPG model had not been used to solve the problems of residential demand response, especially not to be used in generating retail prices for the service provider.

3. Research on Demand Response Model

When there is a power shortage in the real-time balance market, the power generator increases the output, and the system operator bears the cost of system adjustment. Therefore, by introducing demand response into real-time market balance, system operators subjectively design reasonable demand response strategies for various demand responses, guide demand-side responses, and reduce the system adjustment costs. To make reasonable demand response strategies, it is necessary to consider the response characteristics of different market participants and establish the demand response model which is showed in Figure 1. It consists of a power grid company, service provider, customers, and other market participants.
The power grid company usually formulates the wholesale price and supplies the power to the service provider. Power grid companies would evaluate the demand from a large amount of metering and load forecasting data. Because of the electric reformation, some authorities should be delivered to the service provider such as the price ensure of electricity.
The service provider is between the power grid company and customers, operating an incentive-based demand response program with its customers that encourages the customers to sell demand reduction in exchange for a lower retail price. Here we give the contact between retail price and customers’ power reducing ΔEn,t which is described as follows:
Δ E n , t = E n , t ξ t P n , t r e t a i l P min P min
In Equation (1), En,t indicates the energy demand of customer n at time t, and ξt is the elasticity coefficient at time t that denotes the ratio of energy demand change to the retail price. In the paper, Pn,tretail is the retail price at time t for customer n, Pmin is set as the wholesale price which usually is fixed over a period and should be announced by the power grid company.
Customers are enrolled in the demand response program, including large industrial users, load aggregators, end-users, and other market participants. When they are informed of the retail price by the service provider, customers try to maximize their benefits by decreasing their power consumption to get a lower retail price. the electricity benefit is the sum of the electricity costs for actual power consumption which is described as follows:
B b e n e f i t = t = 1 T ( P n , t r e t a i l - P t w h o l e s a l e ) Δ E n , t
In Equation (2), Ptwholesale is the wholesale price, Pn,tretail is the retail price.
Otherwise, a dissatisfaction cost function is used in this paper which is described as follows:
φ n , t ( Δ E n , t ) = μ n 2 ( Δ E n , t ) 2 + ω n ( Δ E n , t )
In Equation (3), μn and ωn are customer-dependent parameters, where μn is a customer label that varies among different power usage. ωn is an auxiliary coefficient of the dissatisfaction cost function, and the details are described in [46].

4. Residual Demand Response Strategy Based on Deep Deterministic Policy Gradient

4.1. The Optimal Model of Power Consumption Strategy

An actual situation that the electric operators often confront is the task of regulating the load. If the power suppliers cannot satisfy the power demand, it will cause economic losses and affect economic developments. In the past, regulating the output of generating units is the only effective approach to address this issue. Once the output limit is exceeded, the normal power loads of customers in the country and other unimportant loads would be forced to cut off. This simple strategy seriously affects the power quality and satisfaction of customers. With the development of China’s economy and the continuous improvement of people’s living standards, the frequency of power blackout has become an important quality index to measure the level of power consumption. The simple load shedding strategy in the past can no longer meet the current demand. Therefore, it is an ideal choice to use incentives to guide customers to actively participate in the load regulation with the participation of the service provider.
As shown in Figure 2, the power grid company is making the wholesale price based on a load forecasting curve and related policy. Instead, the peak shaving and valley filling demand should relay to service providers. Then, the service provider would make the retail price for customers. It is worth noting that both the customer and the service provider need to survive by maximizing their benefits, so retail price making is an optimization question. In other words, the service provider should economically formulate a reasonable retail price to realize the load regulation and maximize the benefits. Meanwhile, customers hope that costs could be minimized according to the existing retail price and the adjustable load.
Combined with the descriptions in Section 3, several settings about this model can be obtained:
(1)
At any time, the total amount to be regulated in load regulation should be less than the adjustable load of customers and the service provider hope the rate of actual power consumption and adjustable load will be greater.
(2)
We set the retail price in a range that is greater than or equal to the wholesale price, and less than or equal to the twice wholesale price. If the retail price is lower than the wholesale price, or the retail price is too higher that customers will no longer wish to buy the electricity, and the survival of service providers will be the biggest problem.
(3)
From the perspective of residential electricity demand, it can be assumed that the total amount of consumers who wish to actively participate in load regulation is no more than 50% of their normal load. The number of consumers who actively participate in demand response would be influenced by the retail price.

4.2. Deep Deterministic Policy Gradient Algorithm

Our proposed DDPG is categorized in the deep reinforcement learning area, depending on actor-critical framework and DQN algorithm. In both actor and critical, there are target net and eval net as the baseline for respective sections. In the training process, only the parameters of the estimation network need to be updated, and the parameters of the target network are directly copied by the estimation network in each loop, as shown in Figure 3. The entire process of DDPG is illustrated in Table 2. At first, network parameters are required to be initialized. Then the related actions are activated, agents will be responded to in the corresponding states, and the instant rewards could be acquired.
Loss function with critical network learning as follows:
y = r + γ max a Q ¯ * ( s , a ) L ( θ ) = E s , a , r , s [ ( Q * ( s , a | θ ) y ) 2 ]
In Equation(4), where network hyper-parameters are estimated according to the critical value. a represents the action from the network estimated by the actor, and y is the Q value of the target network. Because the deterministic strategy is adopted, the greedy algorithm is no longer used to select the action when calculating the target Q value. In general, the training of critical estimation network is still derived from the target Q value and the square loss of the estimated Q value. The estimated Q value is obtained by inputting the current state s and the action a of the actor estimation network into the critical estimation network, while the target Q value is acquired according to the reward R. The Q value that comes from the next state and the action from the actor target network with the discount factor, then the sum of sub-addition is obtained.
The actor network updates the parameters based on the deterministic strategy according to Equation (5).
J ( θ ) = E s D [ θ μ θ ( a | s ) a Q μ ( s , a ) | a = μ θ ( s ) ]
Assuming that the action space is continuous, two different actions are output for the same state. Two feedback values are obtained, based on the state estimation network. More rewards can be acquired by executing actions. The idea of a strategy gradient is to ensure an appropriate probability for obtaining a larger Q value.

4.3. Key Parts of the Model

State space, action space, and reward function are three key parts of the model which directly affect performances of optimal strategy based on DDPG.

4.3.1. State Space

State space S is the observation of the environment which could be changed by the action. In the optimal model, the parameters ΔEn,t are illustrated as Equation (6) which is the demand reduction offered by.
s S ( F t ( l o a d ) , P t w h o l e s a l e , Δ E n , t )
In Equation(6), Ft(load) is the forecasting data for total load, Ptwholesale is the wholesale price formulated by the power grid, ΔEn,t is the actual power consumption of customer n at time t.

4.3.2. Action Space

Action space describes the collection of behaviors and strategies which would change the status of the environment. The target of the optimal model is to improve the devotion of customers to increase the amount of active participation for load regulation. In this paper, the retail price Pn,tretail is chosen as the action behavior which can be described in Equation (7).
a A ( P n , t r e t a i l )

4.3.3. Design of Reward Function

Design of the reward function is an important task for deep reinforcement learning. In this model, the reward system is described in Equation (8) and divided into three parts.
r ( s n , t , a n , t ) R ( Δ E n , t , P n , t retail )
Part one:
R 1 = + 100 , s u c c e s s 1000 , f a i l
In Equation (9), if the retail price for the service provider is during the [Pmin, Pmax] and the benefits of the demand response outweigh the costs, the strategy could be regarded as successful. The reward is increased by 100; otherwise, it should be regarded as failure and the reward is reduced by 1000. In the paper, we set Pmin to be the wholesale price and the Pmax to be twice the wholesale price.
Part two:
R 2 = n = 1 N t = 1 T ( P n , t r e t a i l P t w h o l e s a l e ) Δ E n , h
In Equation (10), the reward function is the sum of economic benefits for the service provider. It includes the electricity sales and benefits for completing the demand response task, which usually is provided by the power grid company or government.
Part three:
R 3 = t = 1 T [ ρ ( P n , t r e t a i l - P t w h o l e s a l e ) Δ E n , t ( 1 ρ ) φ n , t ( Δ E n , t ) ]
In Equation (11), the reward function is the sum of the cost for customers. It includes the electricity cost and the dissatisfaction cost as in Equation (3). The parameter ρ is the weight that describes the classification for customers. The value of ρ is greater, and the customer should pay more attention to the benefit for demand response. Otherwise, the customers should pay more attention to their discomfort when their electricity consumption is reduced.

5. Result Analysis

5.1. Peak-Valley Time of Use (TOU) Price

In this paper, a peak-valley TOU price is used for industrial and commercial electricity consumption as shown in Figure 4. The valley period is from 22:00 p.m. to 8:00 a.m. of the next day, the normal period is from 12:00 a.m. to 18:00 p.m., and the peak period is from 8:00 a.m. to 12:00 a.m. and from 18:00 p.m. to 22:00 p.m.

5.2. Data Set

We chose 867 real customers’ daily power usage data in China between 1st March 2016 and 31st March 2016. Customers were divided into four categories according to daily average electricity consumption, including less than 10 kWh, between 10 kWh and 20 kWh, between 20 kWh and 50 kWh, and between 50 kWh and 100 kWh as shown in Figure 5.
The data set consists of the metering data from a widely installed smart meter. Considering the need for normal power consumption, the adjustable load to be used in demand response cannot be the whole load of customers. In the paper, we divided it into two parts: fixed load and adjustable load power consumption.
Fixed load is a basic electricity usage that does not decrease with the price adjustment. In fact, the power grid company and service provider use at least a month of the power consumption data in the past to ensure and evaluate the fixed load of customers. Unlike fixed load, the adjustable load is affected by electricity price. Usually, we regard the air conditioner and heater as the mainly adjustable load. The work about how to ensure the fixed load and adjustable load has already been researched in many existing papers [47]. The detailed discussion and analysis of the divided method will not be presented in this paper.

5.3. Model Training

In the paper, Q-learning, DQN and DDPG models were training to generate the retail price. As the training processes were similar, the loss value is decreasing to keep stable and accumulative rewards keep increasing until they converged. It should repeat training processes many times to get a good result. A typical training process for DRL model is as shown in Figure 6, which is the model used to generate the retail price based on DDPG.

5.4. Analysis of Retail Price Strategy

The comparison between different retail price curves by using Q-learning, DQN, DDPG models, and peak-valley TOU price are shown in Figure 7, and the service provider’s profit and customer’s cost for four customer groups with the unit “¥“ which is China’s currency yuan were shown in Table 3 based on different RL models.
Because of the range limit, it can be seen that the retail price converges to the value about twice the wholesale price, and the DDPG model has a higher and more stable performance. The retail prices generated by the DDPG model fluctuate within 5% while the retail prices generated by other algorithms fluctuate about 10% or 15%.
As the execution effect of the DDPG model, the actual power consumption is shown in Figure 8 compared with the adjustable load for different groups of customers. The DDPG model could give a suitable retail price for guiding customers to actively participate in the 85–95% percent of demand response programs.

6. Conclusions

With the development of new electric power system reform in China, traditional centralized and unified planning management mode of power grid companies has struggled to meet the needs of development. How to strengthen the interaction between the customers and realize the further optimization of resource allocation have become key research directions. However, the interaction pattern is too complex to build a mathematical or physical model which could completely describe the process. DRL have attracted a great deal of attentions in recent years because of its advantages as an optimal model that can study the interactive experience in a model-free environment. A deep deterministic policy gradient model is proposed to automatically set retail prices for service providers in this paper, which could also realize the maximization of both power sales and power consumption. Obviously, the DDPG model is better for the deterministic policy when compared with Q-learning and DQN algorithms and has improved the training stability and convergence efficiency.
For future studies, further optimization of teamwork in the retail price strategy should guide the customers to actively participate. There are several future works needed:
(1)
In this paper, the optimal model of retail price generation strategy depends on the electricity consumption every 15 min; a shorter time scale should be introduced in the future, such as one minute or in seconds.
(2)
Determine if DRL model is suitable to deal with uncertain interaction process, and how to design the reward function has always been an important issue in modeling.
(3)
Present research focuses on a simulation system that consists of simplified parts. It could be applied in a real situation in the future in the process of electric power system reform in China.

Author Contributions

Conceptualization, K.W.; Data curation, C.D.; Funding acquisition, C.D.; Methodology, C.D.; Software, C.D.; Writing—original draft, C.D.; Writing—review & editing, C.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. State Council of the people’s Republic of China. Some Opinions on Further Deepening the Reform of Power System; State Council of the People’s Republic of China: Beijing, China, 2015. [Google Scholar]
  2. Li, Y.; Huang, L.; Cui, Q.; Shu, J. A Bi-level optimal configuration of combined cooling, heating, and power system considering demand side management. In Proceedings of the 2020 IEEE 4th Conference on Energy Internet and Energy System Integration: Connecting the Grids Towards a Low-Carbon High-Efficiency Energy System, EI2 2020, Wuhan, China, 30 October–1 November 2020; pp. 2064–2070. [Google Scholar]
  3. Li, Y.; Yao, J.; Yong, T.; Ju, P.; Yang, S.; Shi, X. Estimation approach to aggregated power and response potential of residential thermostatically controlled loads. Proc. CSEE 2017, 37, 5519–5528. [Google Scholar]
  4. Ren, Y.; Zhou, M.; Li, G. Bi-level model of electricity procurement and sale strategies for electricity retailers considering users’ demand response. Autom. Electr. Power Syst. 2017, 41, 30–36. [Google Scholar]
  5. Gu, W.; Ren, J.Y.; Gao, J.; Gao, F.; Song, X. Optimal dispatching model of electricity retailers considering distributed generator and adjustable load. Autom. Electr. Power Syst. 2017, 41, 37–44. [Google Scholar]
  6. Hu, Q.; Li, F.; Fang, X.; Bai, L. A framework of residential demand aggregation with financial incentives. IEEE Trans. Smart Grid 2018, 9, 497–505. [Google Scholar] [CrossRef]
  7. Wang, Z.; Paranjape, R. Optimal residential demand response for multiple heterogeneous homes with real-time price prediction in a multiagent framework. IEEE Trans. Smart Grid 2017, 8, 1173–1184. [Google Scholar] [CrossRef]
  8. Liu, X.; Gao, B.; Luo, J.; Tang, Y. Non-cooperative game based hierarchical dispatch model of residential loads. Autom. Electr. Power Syst. 2017, 41, 54–60. [Google Scholar]
  9. Jun, L.; Wenhao, P.; Yanping, Z.H. A scheduling strategy based on particle correction for intelligent power demand response. Power Syst. Technol. 2017, 41, 2370–2377. [Google Scholar]
  10. Yu, M.; Hong, S.H. Incentive-based demand response considering hierarchical electricity market: A Stackelberg game approach. Appl. Energy 2017, 203, 267–279. [Google Scholar] [CrossRef]
  11. De Paola, A.; Angeli, D.; Strbac, G. Price-based schemes for distributed coordination of flexible demand in the electricity market. IEEE Trans. Smart Grid 2017, 8, 3104–3116. [Google Scholar] [CrossRef]
  12. Safdarian, A.; Fotuhi-Firuzabad, M.; Lehtonen, M. Optimal residential load management in smart grids: A decentralized framework. IEEE Trans. Smart Grid 2016, 7, 1836–1845. [Google Scholar] [CrossRef]
  13. Joe-Wong, C.; Sen, S.; Ha, S.; Chiang, M. Optimized day-ahead pricing for smart grids with device-specific scheduling flexibility. IEEE J. Sel. Areas Commun. 2012, 30, 1075–1085. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Kong, X.; Sun, B.; Wang, J. Multi-Time Scale Home Energy Management Strategy Based on Electricity Demand Response. Power Syst. Technol. 2018, 42, 1811–1818. [Google Scholar]
  15. Nunna, H.K.; Doolla, S. Demand response in smart distribution system with multiple microgrids. IEEE Trans. Smart Grid 2012, 3, 1641–1649. [Google Scholar] [CrossRef]
  16. Yu, D.; Xu, X.; Dong, M.; Nojavan, S.; Jermsittiparsert, K.; Abdollahi, A.; Aalami, H.A.; Pashaei-Didani, H. Modeling and prioritizing dynamic demand response programs in the electricity markets. Sustain. Cities Soc. 2020, 53, 101921. [Google Scholar] [CrossRef]
  17. Tang, R.; Wang, S.; Li, H. Game theory based interactive demand side management responding to dynamic pricing in price-based demand response of smart grids. Appl. Energy 2019, 250, 118–130. [Google Scholar] [CrossRef]
  18. Chen, T.; Alsafasfeh, Q.; Pourbabak, H.; Su, W. The next-generation U.S. retail electricity market with customers and prosumers-A bibliographical survey. Energies 2018, 11, 8. [Google Scholar] [CrossRef] [Green Version]
  19. Yang, H.; Wang, L.; Ma, Y.; Zhang, D.; WU, H. Optimization strategy of price-based demand response considering the bidirectional feedback effect. IET Gener. Transm. Distrib. 2021. [Google Scholar] [CrossRef]
  20. Tan, Z.F.; Chen, G.J.; Zhao, J.B.; Hou, J.C.; Jiang, H.Y. Optimization model for designing peak-valley time-of-use power price of generation side and sale side at the direction of energy conservation dispatch. Proc. Chin. Soc. Electr. Eng. 2009, 29, 55–62. [Google Scholar]
  21. Faddel, S.; Mohammed Osama, A. Automated distributed electric vehicle controller for residential demand side management. IEEE Trans. Ind. Appl. 2019, 55, 16–25. [Google Scholar] [CrossRef]
  22. Rigodanzo, J.; da Rosa Abaide, A.; Garcia, V.J.; da Silva, L.N.; Hammerschmitt, B.K.; Bibiano, L.M. Residential Consumer Satisfaction Considering Tariff Variation Based on a Fuzzy Model. In Proceedings of the 2019 IEEE PES Conference on Innovative Smart Grid Technologies, ISGT Latin America 2019, Gramado, Brazil, 15–18 September 2019. [Google Scholar]
  23. Dehnavi, E.; Abdi, H. Optimal pricing in time of use demand response by integrating with dynamic economic dispatch problem. Energy 2016, 109, 1086–1094. [Google Scholar] [CrossRef]
  24. Cui, Q.; Wang, X.; Wang, W. Stagger Peak Electricity Price for Heavy Energy-Consuming Enterprises Considering Improvement of Wind Power Accommodation. Power Syst. Technol. 2015, 39, 946–952. [Google Scholar]
  25. Yang, H.; Wang, L.; Gong, Z.; Zhang, X. TOU electricity price optimization model for multi-customer considering the PV ratio coefficient. In Proceedings of the 2019 4th International Conference on Intelligent Green Building and Smart Grid, Hubei, China, 6–9 September 2019; pp. 345–348. [Google Scholar]
  26. Ding, Y.; Xu, Q.; Xia, Y.; Zhao, J.; Yuan, X.; Yin, J. Optimal dispatching strategy for user-side integrated energy system considering multiservice of energy storage. Int. J. Electr. Power Energy Syst. 2021, 129, 106810. [Google Scholar] [CrossRef]
  27. Wu, L.; Lee, S. A deep learning-based strategy to the energy management-advice for time-of-use rate of household electricity consumption. J. Internet Technol. 2020, 21, 305–311. [Google Scholar]
  28. Li, Q.; Li, J.; Huang, Z.; Fan, F.; Teng, W. A Real-Time Electricity Price Decision Model for Demand Side Management in Wind Power Heating Mode. Front. Energy Res. 2021, 9, 16. [Google Scholar] [CrossRef]
  29. Deng, Z.; Wang, B.; Xu, Y.; Xu, T.; Liu, C.; Zhu, Z. Multi-scale convolutional neural network with time-cognition for multi-step short-term load forecasting. IEEE Access 2019, 7, 88058–88071. [Google Scholar] [CrossRef]
  30. Chang, Z.; Zhang, Y.; Chen, W. Electricity price prediction based on hybrid model of adam optimized LSTM neural network and wavelet transform. Energy 2019, 187, 115804. [Google Scholar] [CrossRef]
  31. Deng, Z.; Liu, C.; Zhu, Z. Inter-hours rolling scheduling of behind-the-meter storage operating systems using electricity price forecasting based on deep convolutional neural network. Int. J. Electr. Power Energy Syst. 2021, 125, 106499. [Google Scholar] [CrossRef]
  32. Nguyen, N.D.; Nguyen, T.T.; Vamplew, P.; Dazeley, R.; Nahavandi, S. A Prioritized objective actor-critic method for deep reinforcement learning. Neural Comput. Appl. 2021, 1–15, in press. [Google Scholar] [CrossRef]
  33. Zhang, D.X.; Han, X.Q.; Deng, C.Y. Review on the research and practice of deep learning and reinforcement learning in smart grids. CSEE J. Power Energy Syst. 2018, 4, 362–370. [Google Scholar] [CrossRef]
  34. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  35. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef] [PubMed]
  36. Gu, S.; Holly, E.; Lillicrap, T.; Levine, S. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates[C]. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Marina Bay Sands, Singapore, 29 May–3 June 2017; pp. 1–9. [Google Scholar]
  37. Burda, Y.; Edwards, H.; Pathak, D.; Storkey, A.; Darrell, T.; Efros, A.A. Large-Scale Study of Curiosity-Driven Learning. In Proceedings of the 7th International Conference on Learning Representations, ICLR, New Orleans, LA, USA, 6–9 May 2019; pp. 1–15. [Google Scholar]
  38. Jang, E.; Devin, C.; Vanhoucke, V.; Levine, S. Grasp2Vec: Learning Object Representations from Self-Supervised Grasping. arXiv 2018, arXiv:1811.06964. [Google Scholar]
  39. Luo, W.; Sun, P.; Zhong, F.; Liu, W.; Zhang, T.; Wang, Y. End-to-end active object tracking and its real-world deployment via reinforcement learning. J. Latex Class Files 2015, 14, 1–16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Fang, K.; Zhu, Y.; Garg, A.; Kurenkov, A.; Mehta, V.; Fei-Fei, L.; Savarese, S. Learning task-oriented grasping for tool manipulation from simulated self-supervision. Int. J. Robot. Res. 2020, 39, 202–216. [Google Scholar] [CrossRef] [Green Version]
  41. Liu, J.; Gao, F.; Luo, X. Survey of deep reinforcement learning based on value function and policy gradient. Chin. J. Comput. 2019, 42, 1406–1438. [Google Scholar]
  42. Antonopoulos, I.; Robu, V.; Couraud, B.; Kirli, D.; Norbu, S.; Kiprakis, A.; Flynn, D.; Elizondo-Gonzalez, S.; Wattam, S. Artificial intelligence and machine learning approaches to energy demand-side response: A systematic review. Renew. Sustain. Energy Rev. 2020, 130, 109899. [Google Scholar] [CrossRef]
  43. Vázquez-Canteli, J.R.; Nagy, Z. Reinforcement learning for demand response: A review of algorithms and modeling techniques. Appl. Energy 2019, 235, 1072–1089. [Google Scholar] [CrossRef]
  44. Zhang, Z.; Zhang, D.; Qiu, R.C. Deep reinforcement learning for power system: An overview. CSEE J. Power Energy Syst. 2019, 6, 213–225. [Google Scholar]
  45. Liu, Y.; Zhang, D.; Gooi, H.B. Data-driven decision-making strategies for electricity retailers: Deep reinforcement learning approach. CSEE J. Power Energy Syst. 2020. [Google Scholar] [CrossRef]
  46. Renzhi, L.; Seung, H.H. Incentive-based demand response for smart grid with reinforcement learning and deep neural network. Appl. Energy 2019, 236, 937–949. [Google Scholar]
  47. Asadinejad, A.; Rahimpour, A.; Tomsovic, K.; Qi, H.; Chen, C.F. Evaluation of residential customer elasticity for incentive based demand response programs. Electr. Power Syst. Res. 2018, 158, 26–36. [Google Scholar] [CrossRef]
Figure 1. Power trading model considering the service provider.
Figure 1. Power trading model considering the service provider.
Processes 09 00660 g001
Figure 2. The optimal model of power consumption strategy based on DDPG.
Figure 2. The optimal model of power consumption strategy based on DDPG.
Processes 09 00660 g002
Figure 3. The principle of network parameters updating for DDPG algorithm.
Figure 3. The principle of network parameters updating for DDPG algorithm.
Processes 09 00660 g003
Figure 4. Peak-valley TOU price curve of typical province.
Figure 4. Peak-valley TOU price curve of typical province.
Processes 09 00660 g004
Figure 5. Electricity consumption curve of typical customers.
Figure 5. Electricity consumption curve of typical customers.
Processes 09 00660 g005aProcesses 09 00660 g005b
Figure 6. Training curve of optimal model.
Figure 6. Training curve of optimal model.
Processes 09 00660 g006
Figure 7. The retail price strategies based on different models.
Figure 7. The retail price strategies based on different models.
Processes 09 00660 g007
Figure 8. The execution effect of DDPG model on four types of customers.
Figure 8. The execution effect of DDPG model on four types of customers.
Processes 09 00660 g008aProcesses 09 00660 g008b
Table 1. The classification for typical reinforcement learning (RL) and deep reinforcement learning (DRL) models.
Table 1. The classification for typical reinforcement learning (RL) and deep reinforcement learning (DRL) models.
ClassificationReinforcement LearningDeep Reinforcement Learning
Based on value function Q-learningDeep Q Network, DQN
Sarsa
Based on policy gradientReinforce algorithmDeterministic policy gradient, DPG
Deep deterministic policy gradient, DDPG
actor–critic, ACTrust region policy optimization, TRPO
Asynchronous advantage actor-critic, A3C
Distributed proximal policy optimization, DPPO
Table 2. The DDPG Algorithm.
Table 2. The DDPG Algorithm.
DDPG Algorithm
1 Random   initialize   network   parameters   of   neural   network   θ Q   and   θ C
2    Initialize   target   network   parameters   θ Q θ Q   and   θ μ θ μ
3   Initialize playback unit R
4   For episode = 1, max_episode do:
5     Initialize   the   random   noise   N for the action to improve the exploration rate
6    Read initialization status S1
7    For t = 1, T do:
8      Select   action   according   to   the   current   strategy   and   random   noise   a t = μ s t θ μ + N t
9      Executive   action   a t ,   get   reward   r t and   new   state   s t + 1
10      Save   the   action   state   sequence   s t , a t , r t , s t + 1 to the playback unit R
11      Randomly   sampled   the   number   of   action   state   sequences   with   the   same   batch   size   from   the   revisit   unit   s t , a t , r t , s t + 1
12      Set   y i = r i + γ Q s i + 1 , μ s i + 1 θ μ θ Q
13      Update   critical   network   parameters   according   to   LOSS :   L o s s = 1 N i y i Q s i , a i θ Q 2
14     Update actor-network parameters according to the policy gradient:
      θ μ J 1 N i a Q s , a θ Q s = s i , a = μ s i θ μ μ s θ μ s i
15     Update target network parameters:
      θ μ τ θ μ + 1 τ θ μ θ Q τ θ Q + 1 τ θ Q
16   End
17End
Table 3. The Comparison of Different DRL Models.
Table 3. The Comparison of Different DRL Models.
Q-LearningDQN DDPG
Group 1Service provider’s profit¥ 96.12¥ 103.24¥ 242.53
Customer’s cost¥ −1.11¥ −3.26¥ 10.31
Group 2Service provider’s profit¥ 80.22¥ 125.0¥ 260.41
Customer’s cost¥ −2.32¥ −1.21¥ 20.76
Group 3Service provider’s profit¥ 108.31¥ 150.51¥ 272.08
Customer’s cost¥ −1.54¥ −2.5¥ 22.43
Group 4Service provider’s profit¥ 96.31¥ 135.51¥ 226.08
Customer’s cost¥ −1.50¥ −2.42¥ 13.41
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Deng, C.; Wu, K. Residential Demand Response Strategy Based on Deep Deterministic Policy Gradient. Processes 2021, 9, 660. https://doi.org/10.3390/pr9040660

AMA Style

Deng C, Wu K. Residential Demand Response Strategy Based on Deep Deterministic Policy Gradient. Processes. 2021; 9(4):660. https://doi.org/10.3390/pr9040660

Chicago/Turabian Style

Deng, Chunyu, and Kehe Wu. 2021. "Residential Demand Response Strategy Based on Deep Deterministic Policy Gradient" Processes 9, no. 4: 660. https://doi.org/10.3390/pr9040660

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop