Journal of Modern Power Systems and Clean Energy

ISSN 2196-5625 CN 32-1884/TK

网刊加载中。。。

使用Chrome浏览器效果最佳,继续浏览,你可能不会看到最佳的展示效果,

确定继续浏览么?

复制成功,请在其他浏览器进行阅读

Electric Vehicle Charging Management Based on Deep Reinforcement Learning  PDF

  • Sichen Li
  • Weihao Hu
  • Di Cao
  • Tomislav Dragičević
  • Qi Huang
  • Zhe Chen
  • Frede Blaabjerg
School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China; Department of Electrical Engineering Center for Electric Power and Energy Smart Electric Components, Technical University of Denmark, Copenhagen, Denmark; Department of Energy Technology, Aalborg University, Aalborg, Denmark

Updated:2022-05-11

DOI:10.35833/MPCE.2020.000460

  • Full Text
  • Figs & Tabs
  • References
  • Authors
  • About
CITE
OUTLINE

Abstract

A time-variable time-of-use electricity price can be used to reduce the charging costs for electric vehicle (EV) owners. Considering the uncertainty of price fluctuation and the randomness of EV owners commuting behavior, we propose a deep reinforcement learning based method for the minimization of individual EV charging cost. The charging problem is first formulated as a Markov decision process (MDP), which has unknown transition probability. A modified long short-term memory (LSTM) neural network is used as the representation layer to extract temporal features from the electricity price signal. The deep deterministic policy gradient (DDPG) algorithm, which has continuous action spaces, is used to solve the MDP. The proposed method can automatically adjust the charging strategy according to electricity price to reduce the charging cost of the EV owner. Several other methods to solve the charging problem are also implemented and quantitatively compared with the proposed method which can reduce the charging cost up to 70.2% compared with other benchmark methods.

I. Introduction

IN recent years, the development of electric vehicles (EVs) has provided a means to reduce air pollution and depletion of conventional carbon energy sources [

1], [2]. Therefore, EV is more suitable for the current environment than the conventional fuel vehicle [3]. In this context, interests in EVs has increased in the scientific community. Most of the existing literature focuses on the social benefit and neglects the benefits to the EV owner [4]-[6]. Considering the economic benefits of EVs to consumers are conducive to promote the transformation of the automobile industry and to increase energy savings and environmental protection benefits. Therefore, we aim to reduce the charging cost of the single EV owner and promote the EV purchase. Since many utility companies utilize the time-of-use electricity price to flatten the demand curve, the charging cost of EV owners can be influenced by the charging/discharging schedules. However, EV charging/discharging schedules face challenges due to the randomness of commuting behavior and electricity price. Thus, a scheduling method that can overcome the challenges is necessary.

Various programming strategies have been proposed to optimize EV charging/discharging schedules, which can be divided into three categories: dynamic programming [

7], [8], non-linear programming [9], and linear programming [10].

A stochastic dynamic programming based method for the scheduling of EV charging is proposed in [

7] to handle the randomness of driving patterns and electricity price. A nonlinear programming based strategy is proposed in [8] to minimize the energy cost of the EV owner. A linear programming method and heuristic algorithm applied from the customer’s perspective to solve determined and dynamic EV charging schedules, respectively [9]. A genetic algorithm and dynamic programming are combined to reduce EV energy consumption [10].

Although programming-based methods capture the law of the interaction between electricity price and charging/discharging behavior to reduce the charging cost of the EV owner, these methods are not always scalable. For a given state, these programming methods require many iterations to obtain the optimal solution. However, the optimization of EV charging cost is a real-time optimization problem. Considering the computation time, the programming-based method is not suitable for the research of this problem [

11].

In recent years, different neural network (NN) based methods have been applied to the research of EV [

12]-[15]. NN can overcome the aforementioned limitations by learning powerful strategies from historical data to address new situations.

The application of NNs in energy management can be divided into two categories: ① NNs assist in making decisions [

16], wherein NNs are utilized to provide the information for other algorithms to manage the energy; ② NNs are directly used for managing the energy [17], [18]. In [17], an energy management controller composed of two NN modules is proposed, and the NN is trained by the results of dynamic programming method to approximate the decision. Similarly, the NN is trained by the Levenberg-Marquardt algorithm in [18]. However, these methods require system information to establish an optimal decision model (DM). In some dynamic random sequential decision problems, these systems are difficult to model.

As a newly developing machine learning, reinforcement learning (RL) can develop an excellent control policy in the absence of initial environment information and the application of RL in decision-making is of great value. In recent literature, RL has solved EV charging schedule problems. Reference [

19] applies the Q-learning algorithm [20] to fast EV charging stations. The results show that the charging cost for the EV owner could be reduced. In [21], RL is used to determine a day-ahead consumption plan for charging a fleet of EVs. Further, the use of the Q-learning algorithm in two different models can reduce the charging costs for the EV owners [22], [23].

The core of Q-learning is an action-value matrix, which is composed of state and action variables whose size determines the complexity of Q-learning. In some low-dimensional state space and discrete action space cases, Q-learning can achieve good performance [

24]. However, many practical applications contain large state and action spaces that create a multi-dimensional action value matrix, making the training difficult. To solve this problem, researchers use an NN approximation method to approximate an action-value matrix in RL. Recently, the DeepMind team successfully solves the problem of non-convergence and instability of an approximate action value function in deep NN [25]-[27] and applies the method to Atari and Go games. Such method of combining deep NN with RL is called deep reinforcement learning (DRL) which has the advantages of overcoming the “dimensional curse”, and does not need system identification steps that may be difficult to obtain in practice. Based on these advantages, the DRL-based methods have been applied to the optimization of wind power forecast uncertainty [28], multi-scenario emergency controller [29], power electronic controller [30], and EV charging scheduling. Specially, [31] considers the randomness of commuting behavior and the uncertainty of electricity price, and the authors apply a naive data-driven deep Q network (DQN) algorithm to obtain a charging strategy without any model information. The results show that the algorithm is effective in reducing the charging cost of the EV owner. However, the discretization of the charging behavior limits the exploration of the action space, which may cause information loss during the training.

We consider an EV charging/discharging model with continuous action spaces, which have a flexible energy management policy, to minimize the charging costs for the EV owner. To overcome the shortcomings of [

31], a DRL-based method that combines the deep deterministic policy gradient (DDPG) algorithm [32] and just another network (JANET) NN [33] to perform real-time optimization of EV charging management is proposed in this paper. The DDPG algorithm is adopted instead of DQN-like algorithms because the discretization of continuous action space causes the loss of significant action information. The JANET NN is used to extract effective temporal information from the electricity price sequence to assist the DDPG algorithm in making decisions. The main contributions of this study are as follows.

1) A DRL-based charging/discharging strategy is proposed for the EV owner. Comparative tests are conducted with different benchmark methods to verify the effectiveness of the proposed method.

2) The novel recurrent neural network (RNN) architecture is used, which is an improved version of LSTM with only the forget gate used to extract the electricity price temporal pattern. A comparative test among different RNN-based feature extraction methods is conducted to demonstrate the impact of the feature extraction ability on the proposed method and to verify the effectiveness of the feature extraction ability of the JANET architecture.

3) Considering the randomness of EV owner’s commuting behavior, the charging/discharging action is decided when the arrival time and departure time of the EV are unknown.

The remainder of this paper is organized in the following structure. In Section II, the single EV charging/discharging scenario is introduced and modeled as a Markov decision process (MDP). The DDPG algorithm and JANET NN are described in Section III. Section IV describes the NN architecture, experimental details, and training process. In Section V, experimental results are presented in detail to demonstrate the effectiveness of the proposed method. Section VI presents comparison results with similar methods and analysis of the simulation results, and Section VII presents the conclusions.

II. Problem Formulation

Assuming that the EV can transmit the power to or receive power from the power grid. The arrival time and plug-in time of EV are tarr and tarr+1 on day 𝒳, respectively. EV departure time is tdep on day 𝒳+1. The episode begins when the EV arrives home on day 𝒳, and ends when the EV leaves home on day 𝒳+1.

In this paper, the charging process is defined as an MDP, which has unknown transition probabilities due to the randomness of EV owner’s commuting behavior and electricity price. This method utilizes the fluctuation in electricity price to minimize the cost. For example, if the EV is charged when electricity price is low and discharged when the electricity price is high, the reduction in charging costs for the EV owner can be achieved. The scenario of this model is shown in Fig. 1, the EV owner has an intelligent charging device (ICD) at home. When the battery is connected to the ICD, the ICD can perform charging/discharging action according to the proposed method. The proposed method needs the real-time remaining capacity of the battery and the previous N-hour electricity price of current time t to make decisions from the EV owner’s perspective.

Fig. 1  Single EV charging management model.

The problem of economic benefits of charging/discharging for the EV owner is modeled as an MDP, which has unknown transition probability with finite time steps. An MDP is a four-tuple 𝒮,  𝒜,  ,  𝒯, where 𝒮 is the state space, 𝒜 is the action space, is the reward function, and 𝒯 is the state transition function.

At time step t, the ICD obtains state st𝒮, which includes the remaining capacity of the battery and the previous N-hour electricity prices of time t. Action at𝒜 is taken, which indicates the charging/discharging power of battery. After the action at is executed, the agent receives an immediate reward rt=(st,at), and the system transfers to a new state st+1=𝒯(st,at). An episode of MDP consists of a finite sequence of time steps, states, actions, rewards, and new states, at the first moment, there is s1, a1, r1, s2; the same to the second moment s2, a2, r2, s3; and at the last moment T, there is sT, aT, rT. The details of MDP formulation are defined as follows.

1) State: at time t, the state of the MDP is represented as st=(Et,Pt-N,Pt-N-1,,Pt-1), where Et is the remaining battery capacity of the EV, and (Pt-N,Pt-N-1,,Pt-1) is the previous N-hour electricity price at time t.

2) Action: at time t, the action is set to be at. The action of MDP is defined as the charging/discharging power, which can be selected continuously in the range -Pch,max-Pch,max, where Pch,max indicates the maximum charging power of EV.

3) Reward function: the reward function (st,at) can be expressed as:

(st,at)=-γPtattarr<t<tdep-p1(Emax-Et)2Et>Emax,tarr<t<tdep-p2(Emin-Et)2Et<Emin,tarr<t<tdep-p3(Emax-Et)2t=tdep (1)

where Emax is the maximum capacity of EV; tarr<t<tdep denotes the time when the EV is at home; t=tdep denotes the time when EV leaves home; Pt is the electricity price at time t; and γ, p1, p2, and p3 are the real-valued coefficients. Therein, these four coefficients are set to ensure that the power demand and economic benefits of the EV owner are satisfied, and the battery runs in a safe working mode.

During the V2G time of EV, -γPtat indicates the charging cost at time t. The two penalty terms -p1(Emax-Et)2 and -p2(Emin-Et)2 are added for safe operation of batteries. -p3(Emax-Et)2 is the penalty term for the EV leaving home without being fully charged. In a real-world scenario, different EV owners have different driving distance demands, some of whom are more concerned with driving distance and others are more concerned with economic benefits. The proposed method considers EV owner’s demand and uses parameter p3 to adjust the characteristics of the model to satisfy different demands, the detailed experiences are shown in Section V.

4) State transition function: the state transition function can be expressed as st+1=𝒯(st,at). In the deterministic part, at only influences Et+1, and the relationship between Et and Et+1 is Et+1=Et+at. In the stochastic part, the transition function, which has unknown transition probability, follows the stochastic conditional probability P(st+1|st,at), which is influenced by the randomness of electricity price and EV owner’s commuting behavior. In a model-based method, it is difficult to model an environment with such a stochastic conditional probability. This paper presents a model-free method to solve this problem by learning the state transition from unlabeled real-world data without designing an environmental dynamic model.

III. Method Introduction

A. RL and EV Charging Strategy

When an agent performs a task, it chooses an action according to policy π to interact with the environment. After it implements the action, a new state is reached and the environment returns a reward to the agent. This process cycles until the agent completes the task well. The objective of RL can be defined as max(R), where R=j=1Tγ(j-1)r(sj,π(sj)), and policy π creates a mapping between the current state and the action to be applied (the action is modeled as a probability distribution). r(sj,π(sj)) is a reward function, T means one episode has T steps, and γ[0,1] is the discount factor used to indicate the importance of future rewards relative to immediate rewards. However, π may be stochastic, which leads to R being stochastic as well. In order to solve the stochastic R, the objective of RL can be defined as max(Eπ[R]).

The action value function is used in RL to improve the policy π to achieve the objective max(Eπ[R]). The action value function Qπ() describes the cumulative expected reward obtained after taking action at at state st, and thereafter using policy π [

32]:

Qπ(st,at)=Ej=tTγ(j-t)r(st,at)|st,at (2)

Its Bellman equation [

34] is:

Qπ(st,at)=Er(st,at),st+1~Er(st,at)+γEat+1~π[Qπ(st+1,at+1)] (3)

where E is the environment.

In this paper, the goal of the ICD is to reduce the charging cost for the EV owner during tarr+1 to tdep. The EV charging scheduling is a sequential decision problem and it is not only influenced by economic benefits of the current time, but also influenced by the economic benefits and the battery energy in the future. As illustrated in (3), the immediate reward of charging/discharging is r(st,at) and γEat+1~π[Qπ(st+1,at+1)] is the future reward.

The proposed method uses a feature analysis model (FAM) to determine the potential patterns from historical electricity price data. Then, RL performs charging/discharging action based on the received features of future electricity price and Et information. Since the agent of the model has continuous action variables, compared with an agent that executes discrete action, the two agents in the same state have a different number of actions that can be selected, which leads to a much larger Q table dimension in the former than in the latter. In the training process, if the Q value of the agent performing continuous actions is calculated, the iterative calculation of the Q table increases dramatically, leading to a time-consuming training process that is difficult to converge [

11]. To avoid such an outcome, we consider an NN approximator parameterized by ω to approximate the action value function [26]:

Qs,a;ωQs,a (4)

B. DDPG Algorithm

The DDPG algorithm is a DRL which is based on (4). The DDPG algorithm consists of two parts, i.e., the critic and the actor parts. The critic part approximates the action value function, and the actor part approximates the strategy function. The connection between the two parts is as follows: the environment provides st to the agent, and the actor part of the agent makes an action at based on st. When the environment receives at, it gives the agent a reward rt and a new st+1. The agent must then update the critic part according to the reward, and then update the actor part in the direction suggested by the critic part. The algorithm moves to the next step and the process continues until a good actor is achieved, which is reflected by a high total reward.

There are four networks included in the DDPG algorithm [

32]: the critic network Qs,a;ω with parameter ω, a copy of the critic network Q's,a;ω' known as the critic target network with parameter ω', the actor network μs;θ with parameter θ, and a copy of the actor network μ's;θ' known as the actor target network with parameter θ'. The two-copy network is used to calculate the target values to improve the stability of the algorithm.

The DDPG algorithm is a deterministic strategy. To find a better strategy, we add Gaussian noise 𝒩 to increase the randomness of the output action in the model.

at=μ(s;θ)+𝒩 (5)

In this algorithm, the loss function is defined as [

32]:

LDDPG=1Nt=1N(yt-Q(st,at;ω))2 (6)
yt=rt+γQ'(st+1,μ'(st+1;θ');ω') (7)

where N is the batch size.

In (6) and (7), the gradient descent method is used to update the parameter ω in the direction of reducing the loss. To update the actor network, the gradient is defined as [

32]:

θμ|st1Nt=1NaQ(s,a;ω)|s=st,a=μ(st)θμ(s;θ)|s=st (8)

In (8), the parameter θ of the strategy is updated in the direction that increases the Q(s,a;ω).

In the DDPG algorithm, a target network parameter updating the method based on the “soft” mode is adopted; the critic/actor target network slowly tracks the critic/actor network parameter. This parameter updating method can significantly increase the stability of learning [

32].

ω'=τω+(1-τ)ω' (9)
θ'=τθ+(1-τ)θ' (10)

where τ1.

C. JANET

Reference [

33] built upon the idea of the gate recurrent unit (GRU) [35] and succeeded in designing the JANET network. The JANET has fewer parameters but performs better in some applications than the standard LSTM model. As shown below, the standard LSTM [36], [37] is defined as:

gt=(Ucht-1+Wcxt+bc) (11)
it=σ(Uiht-1+Wixt+bi) (12)
ft=σ(Ufht-1+Wfxt+bf) (13)
ot=σ(Uoht-1+Woxt+bo) (14)
ct=ft·ct-1+it·gt (15)
ht=ot·(ct) (16)

where gt, it, ft, ot, ct, and ht are the input node, input gate, forget gate, output gate, cell state, and hidden state, respectively; U and W are the matrix weights; bc, bi, bf, and bo are the vectors of biases; is the tanh function; σ is the sigmoid function; and · is the element-wise multiplication operation.

LSTM NN has two features. One is cell state ct, which has a recurrent self-connected edge with a constant weight of 1 to overcome gradient disappearance and gradient explosion [

38]; and the other is three gates it, ft, and ot [39]-[41]. A gate can selectively control the data flow through it. For example, it and ft control the size of data flow into ct, and ot controls the size of data flow into ht. Specifically, gt uses the function to activate the input data xt at the current time step and hidden state ht-1 at the previous time step. it uses the σ function, which can output the values between 0 and 1 to control the data flow from the input node to ct. The σ function is used by ft to control the effect of ct-1, which contains the information of all previous time steps on ct at the current time step. As with it and ft, ot uses the σ function to determine how much (ct) is saved in ht.

The architecture of JANET retains the two features of LSTM but removes it and ot. In addition, although of ht brings the same dynamic output range to each cell, it also causes training difficulties [

37]. Because the vanishing gradient may be deteriorated by the activation of ht [33], the unstable factors in of ht are removed from the JANET architecture.

The proposed method has four JANET layers. The previous 24-hour electricity price data are processed by matrix Win and input into the first JANET layer. Win is the optimization parameter. The features of future electricity price are the outputs at the fourth JANET layer.

Electricity price data are processed before they are input into the JANET cell.

xt=WinPi    i=t-1,t-2,,t-24 (17)

After the data flow into the JANET cell, the hidden state ht1 of the first layer is computed as:

gte=(Uceht-1e+Wcehte-1+bce) (18)
fte=σ(Ufeht-1e+Wfehte-1+bfe) (19)
cte=fte·ct-1e+(1-fte)·gte (20)
hte=cte (21)

where e indicates the eth JANET layer, and there is hte-1=xt at the first layer.

At the fourth layer, the features of future electricity price can be calculated as:

=Woutht4 (22)

where Wout is the optimization parameter. Then, in order to update the parameters of JANET, the loss function can be defined as:

LJANET=1Ni=1N(Pti-i)2 (23)

where Pti is the electricity price of the current time t.

IV. Experimental Settings

A. Deep NN Architecture

As shown in Fig. 2, the Pt-N,Pt-N-1,,Pt-1 are input into the JANET layer to map to . Therein, there is a four-layer JANET network with 50 neurons to each layer. is a not only concatenated with Et, but also with Et and at. Both the actor and critic networks have the same three-layer fully-connected layer with 100 neurons adopted by rectified linear units (ReLU) [

42] at each layer. Lastly, these concatenated information are fed into fully-connected layer of actor and critic network in order to approximate at and Q, respectively.

Fig. 2  DRL method combining DDPG algorithm and JANET NN to perform real-time optimization of EV charging management strategy.

B. Training Process

The training process of the FAM is performed in a supervised manner. The training data contain electricity prices for the first 200 days of 2017 [

43]. In each training iteration, the training data are divided into two parts: the input electricity price and the corresponding desired outputs. The FAM creates a mapping between input previous N-hour and output predicted electricity prices during the training process. It adjusts the parameters of the NN at each iteration to minimize the differences between the electricity price of the FAM output and the desired electricity price.

After the training of FAM is completed, the training of DM can be implemented based on the FAM output. The training process and the main parameters of the DDPG are shown in Algorithm1 and Table I, respectively.

TABLE I  Parameters of DDPG
ParameterValue
Reward discount factor γ 1
Capacity of memory D 6×104
Learning rate of actor 4×10-6
Learning rate of critic 8×10-6
Batch size 256
Training epoch 2.1×105
Number of hidden layer 3
Number of hidden units per layer 100
Nonlinearity of hidden layer ReLU

Algorithm 1  : the training of DDPG of EV charging model

1. Initialize the hyper-parameter

2. Initialize the M-sized replay buffer D

3. Initialize weights ω, ω', θ and θ'

4. for episode ranging from 1 to M do

5.    Randomly choose X day from training data

6.    Randomly choose tarr, tdep, and battery energy at time tarr

7.    for tarr to tdep do

8.       Extract FAM output features after receiving previous

         24-hour electricity price from st

9. Concatenate features with battery energy as C

10.      Choose action at=μC;θ+N

11.      Enter the action at and state st into the environment to

          obtain the reward rt and the next state st+1

12.      Store the transition (st, at, rt, st+1) in D

13.      if D is full then

14.         Randomly sample Z-sized transitions (st, at, rt, st+1)

15.         if episode is even then

16.             Update ω with (6) and (7)

17.             Update θ with (8)

18.             Update ω' with (9)

19.             Update θ' with (10)

20.         end if

21.      end if

22.   end for

23. end for

At the beginning of the DM training process, the replay buffer 𝒟 and four NNs are established. The purpose of establishing replay buffer 𝒟 [

25], the critic target network Q', and the actor target network μ' [26] is to break the temporal pattern between the training data to increase the robustness of training. After initializing, the proposed DRL method is trained for 210000 episodes to learn the optimal EV charging/discharging strategy. We use real-time electricity price data [43] of zone COMED of PJM, USA to train and test the proposed method. The data are divided into training data and test data. The training data contain the data from the first 200 days of 2017, and the test data are from 201-300 days of 2017. 𝒳 randomly chooses from the first 200 days of 2017, and an episode begins at tarr and ends at tdep, thus the length of episode is not fixed. The FAM parameters are loaded at the beginning of training, and the FAM outputs the features of future price after receiving the historical price. As shown in line 9 of Algorithm 1, the DM makes decision based on the FAM output and battery energy. Memory capacity is set to be 60000, and the learning begins when the replay memory is full. It is important for the agent to explore the environment. Therefore, the proposed method uses an interval episode learning method to train the model.

C. Training and Practice Workflow

The complete workflow of the proposed method is shown in Fig. 3, where Cell is the network parameter. The workflow can be divided into the training phase with two steps and the practice phase with one step. In the training phase, the training FAM is first executed. The previous 24-hour electricity price data of current time t are input to the FAM to map to . The updating process of FAM is actually a supervised learning updating process, thus the parameters of FAM can be updated in the direction of minimizing the (23) after is obtained. The second step of the training phase is training DM. As shown in Fig. 3, the DM takes Et and as inputs. The DM belongs to DRL, so it needs to interact with the environment to explore and exploit, and update its parameters by (8). When the DM finishes its training, the training phase is completed and it moves to the practice phase. In the practice phase, the EV will equip the ICD and connect to the power grid. In this step, Et and previous 24-hour electricity price data are input to the ICD, where Et will be directly input to DM, and the previous 24-hour electricity price data will be processed by FAM and then input into DM. It should be noted that DM in practice phase only needs actor part to work, while critic part of DM only needs to be used in training phase.

Fig. 3  Complete workflow of proposed method.

D. Experimental Details

To show the randomness of commuting behavior, Etarr, tarr, and tdep are generated randomly. Et of EV when arriving home follows a normal distribution Nμ,σ2, where μ=0.45, and σ=0.1. tarr and tdep follow a uniform distribution and are sampled from the sets of {15, 16, 17, 18, 19, 20} and {6, 7, 8, 9, 10, 11}, respectively.

We use a FIAT 500e with battery storage of Emax=24  kWh and Emin=1  kWh in the experiments. The maximum charging power and discharging power of the battery are assumed to be 6 kW and -6 kW, respectively.

The experimental environment is implemented in Python using Tensorflow. The experimental workstation is a computer with an Intel Core i5-6300HQ and a NVIDIA GTX960M GPU.

V. Experimental Results

A. Case 1

1) Training results: the training data contain the data for the first 200 days of 2017 [

43]. The training accuracy of the FAM is shown in Fig. 4(a). It can be observed from the Fig. 4(a) that the prediction error of NN is gradually decreasing with the advance of training which demonstrates that the FAM can learn the pattern of the training data. In supervised learning, the training accuracy does not fully represent the validity of the model. Thus, the effectiveness of the FAM is further discussed in case 2. The training process of the DM is shown in Fig. 4(b). It is observed in Fig. 4(b) that the cumulative reward begins to increase sharply near 4200 episodes, and slowly increases until 210000 episodes. Figure 4(b) shows that the proposed DRL-based method can learn an valid policy to obtain a high cumulative reward.

Fig. 4  Training results of FAM and DM. (a) Training accuracy of FAM. (b) Cumulative reward (average value) of each episode during DM training process.

2) Model performance: the test data are from 201-300 days of 2017 [

43], which are the total 100-day test data. To demonstrate the performance of the JANET FAM, Fig. 5 shows a comparison of forecasted electricity price and actual electricity price for days 201-230 of 2017. As shown in Fig. 5, the red line is generally similar to the blue line except the very small proportion of very high peak electricity price. The effectiveness of the JANET FAM is described in case 2.

Fig. 5  Comparison of forecasted and actual electricity prices for days 201-230 of 2017.

The electricity price and charging/discharging behavior in four consecutive days are illustrated in Fig. 6 to demonstrate the effectiveness of the proposed method.

Fig. 6  Electricity price and charging/discharging behavior in four consecutive days. (a) 1st day. (b) 2nd day. (c) 3rd day. (d) 4th day.

In Fig. 6(a), the EV owner arrives home at tarr=15 hour and Etarr=13.75  kWh. The V2G time lasts for 19 hours until the EV owner leaves home at tdep=11 hour. In a similar way, for Fig. 6(b)-(d), when tarr=17, 16, 18 hours, Etarr=8.91, 9.04, 10.89 kWh, and tdep=7, 6, 9 hours, respectively. It is observed that when the electricity price is high, the discharging action will be executed, and when the electricity price is low, the charging action will be excuted. In order to further show the performance of the proposed method, Table II illustrates the data of 12 consecutive days in the 100-day test set. Therein, days a to d shown in Table II correspond to Fig. 6(a)-(d). In Table II, 𝒞pro and 𝒞un represent the charging costs per kWh of the proposed method and unmanaged strategy, respectively.

TABLE II  Twelve Consecutive Days of Data in 100-day Test Set
Daytarr (hour)Etarr (kWh)tdep (hour)Etdep (kWh)Cpro($/kWh)Cun($/kWh)
a 15 13.75 11 21.98 -0.0082 0.0489
b 17 8.91 7 21.64 0.0178 0.0284
c 16 9.04 6 21.57 0.0099 0.0364
d 18 10.89 9 21.69 0.0088 0.0392
e 18 9.06 11 21.53 0.0032 0.0494
f 16 9.68 7 21.24 0.0152 0.0316
g 16 5.35 11 21.41 0.0180 0.0374
h 15 14.28 10 21.85 -0.0564 0.0859
i 18 12.80 8 22.12 -0.0058 0.0486
j 20 11.51 10 20.17 0.0296 0.0379
k 19 9.56 8 21.17 0.0176 0.0330
l 17 10.05 9 21.97 0.0213 0.0302

The detailed calculation of the charging cost is presented in the next section. In addition, a trained model can make a decision in 3 ms and it can fully meet the online request.

In a real-world scenario, different people have different driving distances to the individual destination. Those who drive a long distance pay more attention to Etdep than the economic benefits. In contrast, those who drive a short distance pay more attention to the economic benefits than to Etdep. To measure EV owner’s preference, p3 in (1) is introduced, and the two scenarios can be switched as long as p3 is adjusted. Specifically, p3 is set to be 2 for the users with a long driving distance and 1 for the users with a short driving distance. The detailed parameters mentioned in (1) are summarized in Table III. To clearly show the difference in proposed method in two different scenarios, the simulation results are listed in Table IV. Therein, LD_Etdep, SD_Etdep, 𝒞LD, and 𝒞SD represent Etdep in the long-distance driving scenario, Etdep in the short-distance driving scenario, the charging cost in the long-distance driving scenario, and the charging cost in the short-distance driving scenario, respectively. It can be observed from Table IV that the LD_Etdep is closer to Emax than SD_Etdep, and the 𝒞SD is lower than 𝒞LD. The results in Table IV indicate that the proposed method can adaptively adjust different requirements of EV owners by setting different p3.

TABLE III  Parameters of Two Scenarios
Scenarioγp1p2p3
Long-distance driving 7 4 4 2
Short-distance driving 7 4 4 1
TABLE IV  Simulation Results in Twelve Consecutive Days
Daytarr (hour)Etarr(kWh)tdep (hour)LD_Etdep(kWh)SD_Etdep(kWh)CLD $/kWhCSD $/kWh
a 15 13.75 11 23.34 21.98 -0.0051 -0.0082
b 17 8.91 7 23.26 21.64 0.0298 0.0178
c 16 9.04 6 22.93 21.57 0.0242 0.0099
d 18 10.89 9 23.39 21.69 0.0176 0.0088
e 18 9.06 11 23.04 21.53 0.0059 0.0032
f 16 9.68 7 23.02 21.24 0.0277 0.0152
g 16 5.35 11 23.22 21.41 0.0400 0.0180
h 15 14.28 10 23.39 21.85 -0.0662 -0.0564
i 18 12.80 8 23.31 22.12 -0.0019 -0.0058
j 20 11.51 10 21.38 20.17 0.0424 0.0296
k 19 9.56 8 23.22 21.17 0.0294 0.0176
l 17 10.05 9 23.29 21.97 0.0199 0.0213

B. Case 2

The training data and test data mentioned in case 1 are used in this case to investigate the performances of different FAMs and the effect of the FAMs on the DM. To further investigate whether the combination of convolutional NN (CNN) and RNN has a stronger ability to extract the temporal pattern than a single RNN, each RNN adds an additional comparative model that combines the CNN and RNN [

44]. The success of CNN lies in its ability to effectively extract features from the original input data. Therefore, to enhance the feature expression of the original electricity price data, the CNN layer is set before the RNN layer.

The prediction accuracy of different models is tested first. All models have the same parameters and training episodes, as shown in Table V. To measure the performances of these models, the error function is introduced as a metric MSEavergae:

MSEavergae=MSEm=1m1ni=1n(P̂test(i)-Ptest(i))2 (24)
TABLE V  Parameter List of Two Networks
NetworkTraining episodeNumber of units per layerCNN layerRNN layer
CNN + RNN 2000 50 1 3
RNN 2000 50 4

where m is the experiment time; MSE is the mean square error; Ptest is the prediction value; P̂test is the real electricity price; and n is the number of elements in the Ptest set.

The prediction errors of eight models in the 100-day test data are shown in Fig. 7. It is observed that the JANET model demonstrates the best performance and the CNN + bidirectional long short-term memory (BiLSTM) model shows the worst performance of the eight models. In addition, the data of LSTM and GRU models indicate that the CNN + RNN has better ability to extract the temporal pattern in the Ptest set than a single RNN. However, the other data indicate that the RNN network performs better than the CNN + RNN.

Fig. 7  Prediction errors of eight models in 100-day test set.

After studying the accuracy of different FAMs to predict the future electricity price trend, the effect of the different FAMs on the DM is investigated. To visualize the differences among the eight models, the cumulative charging cost of each model in the 100-day test set is subtracted from the charging cost of JANET. The cumulative cost data are obtained by combining the eight RNN models with the same DM and the results are shown in Fig. 8.

Fig. 8  Differences in cumulative charging cost of each model in 100-day test set and JANET model.

Specifically, the economic cost differences of the CNN + JANET, CNN + GRU, GRU, CNN + LSTM, LSTM, BiLSTM, and CNN + BiLSTM models with JANET model are $0.6, $0.88, $1.15, $1.39, $1.56, $1.7, and $2.37 for the same DM, respectively.

The proposed method has two components, the FAM and the DM. Figure 7 shows the ability of different FAMs to extract future electricity price features. Figure 8 shows the effect of different FAMs on the same DM. Figures 7 and 8 indicate that a stronger ability of the FAM to extract features produces a more ideal DM performance. Therefore, the proposed method chooses the JANET NN as the FAM to extract the temporal pattern of electricity price. The data in Figs. 7 and 8 are summarized in Table VI.

TABLE VI  Data from Different FAMs
FAMMSEaverage ($/kWh)Cost difference with JANET ($)
CNN + BiLSTM 0.2845 2.37
BiLSTM 0.2659 1.70
LSTM 0.2566 1.56
CNN + LSTM 0.2400 1.39
GRU 0.2207 1.15
CNN + GRU 0.1981 0.88
CNN + JANET 0.1960 0.60
JANET 0.1853 0

VI. Discussion

We propose a DRL-based method for the charging strategy to reduce the charging cost for the EV owner. The proposed method uses JANET, an improved version of LSTM, as the FAM to extract the variation regularity of electricity price, and applies a DRL algorithm to make decisions based on the extracted features. The proposed method combines the feature extraction ability of deep learning and the decision-making ability of RL, and provides better robustness for the uncertainty of electricity price and EV owner’s commuting behavior.

The research in this paper is similar to [

16] and [31], which focuses on utilizing electricity price fluctuation to reduce the charging cost for the EV owner with a single EV. Although the simulation results of [16] show that the charging behavior can be implemented when electricity prices are low, there is no discharging action when the electricity prices are high. To further reduce the charging cost, discharging behaviors are necessary. The algorithm in [16] is Q-learning, which is difficult to train when facing a multi-demensional action value matrix, and may affect the performance. In addition, the state and action of Q-learning must be discrete variables since the matrix only has the finite size and cannot be generalized. In this way, it may lead to the lack of state and action information and cannot achieve good training results. To avoid the “curse of dimensionality” and “lack of information”, which may cause Q-learning not to work, the main algorithm in [33] is a deep Q-learning network that utilizes NN to approximate the action value matrix. However, simple utilization of a fully-connected layer to handle electricity price may not achieve the desired effect. Considering the strong time characteristic of electricity price, we utilize LSTM-like NN to extract temporal features from the electricity price signal before making decisions.

The results of case 1 show that the proposed method can learn an optimal charging strategy to manage the dynamics of electricity price. Figure 4(b) shows that the value of the reward function proposed in this paper increases with increasing iteration steps until it reaches a convergence value, indicating that the proposed method can learn from the training set to improve the reward function. From Fig. 6, the discharging action is performed at a higher electricity price and the charging action is executed at a lower electricity price, which demonstrates the effectiveness of the reward function and the interpretability of the proposed method. Table II shows more detailed data on the effect of reducing charging cost in the proposed method. From Table IV, the proposed method can switch between the long-distance driving scenario and the short-distance driving scenario as long as p3 is adjusted.

In case 2, the effect of the FAMs on the DM is investigated. Figures 7 and 8 indicate that stronger ability of the FAM to extract features results in more ideal DM performance. Although we focus only on the performances of LSTM-like NNs in feature extraction of electricity price, it has a broader scope. Further research in this area will be conducted in the future.

In order to further verify the effectiveness of the proposed method, different benchmark methods are investigated. The training data and test data of these methods are the same as those in case 1. The proposed method is compared with several baselines as follows.

1) RL-based methods: including DQN charging method in [

31], DQN-with-JANET (DQWJ) charging method, DDPG-with-NN (DWN) charging method, and DDPG-with-CNN + BiLSTM (DWCB) charging method. DWN, DWCB, and the proposed method have the same hyper-parameters and DM. The only difference between them is the FAM. DQN and DQWJ are based on the DQN charging method, the difference between them is whether there is FAM dealing with the uncertainty of electricity price. In addition, the only difference between DQWJ charging method and proposed method is DM.

2) Unmanaged strategy: the unmanaged strategy charges the battery with a maximum power of 6 kW at tarr<t<tdep until the battery storage is full at Emax=24  kWh.

3) Theoretical limit: for the theoretical limit (MATLAB toolbox), tarr, Etarr, tdep, and electricity price are already known before, and a global optimal decision can be made.

Considering the probabilistic events that Etdep is not full at 24 kWh in DQN, DQWJ, DWN, DWCB, and the proposed method, the cumulative charging cost 𝒞 of the these methods can be calculated as:

𝒞=Ptempddep+t=tarrtdepPtat (25)

where ddep=Emax-Etdep; and Ptemp is the first price greater than 0 after tdep. The same rules are also applied in Tables II and IV. For intuitive comparison, the percentage P of cost reduction compared with the unmanaged strategy can be defined as:

P=1-𝒞𝒞un (26)

The cumulative charging costs of all methods in the 100-day test set are shown in Fig. 9 and the detailed data of (25) and (26) are summarized in Table VII.

Fig. 9  Comparison of cumulative charging costs between proposed charging method and four other charging methods in 100-day test set.

TABLE VII  Data of Baselines and Proposed Method
MethodC ($)P (%)
Unmanaged 57.51 0
DQN 41.50 27.84
DQWJ 33.52 41.71
DWN 21.71 62.25
DWCB 19.50 66.09
Proposed 17.13 70.21
Theoretical limit -20.95 136.45

By analyzing the results in Table VII, several conclusions can be drawn. First, the simulation results of DQN and DQWJ indicate that the DM can make a more effective decision after the electricity prices are processed by the FAM. These results show that the randomness of electricity price will affect the training of DM and reduce the performance of scheduling EV charging. Therefore, FAM, a preprocessing module to reduce the randomness of electricity price, is essential to improve the performance of DM. Second, compared with the DQWJ method, the proposed method with a continuous action space learns a better energy management policy to minimize the charging costs for the EV owner. This is understandable since continuous action spaces have a larger solution space to search, providing a good foundation for finding the optimal solution. The DWCB method, with the worst performance of the eight models, has a better performance than that of the DWN method, indicating that an RNN has a stronger ability than a fully-connected NN in smart EV charging management due to the ability of RNN in addressing time series data.

The DRL model of EV charging proposed in this paper can provide customized charging strategies for any specific EV to reduce charging costs. In addition, any deferrable load can use a variant of the proposed method to produce certain economic benefits not limited to EV. However, if a model proposed by large-scale EV owners avoids the peak price and charges at a relatively inexpensive time, the electricity price will rebound due to the economic regulation of the market, which will introduce new uncertainties. In order to avoid introducing the new uncertainties into large-scale optimization, the spatiotemporal pattern based system [

45]-[47] which considers both the temporal and spatial attributes may be one of the research areas. Related literature can refer to [48] and [49]. The relevant research will be further investigated in the future.

VII. Conclusion

The EV is a leading product to drive a new industrial revolution. To promote the transformation of the market from fuel vehicles to EVs, consumer choice is a critical factor. Therefore, it is necessary to develop a strategy for reducing the EV charging cost to increase EV purchasing.

In this context, we propose a DRL-based method that combines the feature extraction ability of deep learning and the decision-making ability of RL for an EV charging strategy that reduces charging cost for the EV owner. The proposed method uses JANET, an improved version of LSTM, as the FAM to extract the variation regularity of electricity price, and applies a DRL algorithm to make decisions based on the extracted features. The simulation results show that the proposed method can reduce the charging cost up to 70.2% compared with other methods.

References

1

Z. Liu, Q. Wu, L. Christensen et al., “Driving pattern analysis of nordic region based on national travel surveys for electric vehicle integration,” Journal of Modern Power Systems and Clean Energy, vol. 3, no. 2, pp. 180-189, Jun. 2015. [Baidu Scholar] 

2

T. Chen, X. Zhang, J. Wang et al., “A review on electric vehicle charging infrastructure development in the UK,” Journal of Modern Power Systems and Clean Energy, vol. 8, no. 2, pp. 193-205, Mar. 2020. [Baidu Scholar] 

3

S. Li, W. Hu, D. Cao et al., “A multi-agent deep reinforcement learning-based approach for the optimization of transformer life using coordinated electric vehicles,” IEEE Transactions on Industrial Informatics. doi: 10.1109/TII.2021.3139650 [Baidu Scholar] 

4

R. Gough, C. Dickerson, P. Rowley et al., “Vehicle-to-grid feasibility: a techno-economic analysis of EV-based energy storage,” Applied Energy, vol. 192, pp. 12-23, Apr. 2017. [Baidu Scholar] 

5

J. Hu, H. Zhou, Y. Li et al., “Multi-time scale energy management strategy of aggregator characterized by photovoltaic generation and electric vehicles,” Journal of Modern Power Systems and Clean Energy, vol. 8, no. 4, pp. 727-736, Jul. 2020. [Baidu Scholar] 

6

H. Patil and V. N. Kalkhambkar, “Grid integration of electric vehicles for economic benefits: a review,” Journal of Modern Power Systems and Clean Energy, vol. 9, no. 1, pp. 13-26, Jan. 2021. [Baidu Scholar] 

7

E. B. Iversen, J. M. Morales, and H. Madsen, “Optimal charging of an electric vehicle using a markov decision process,” Applied Energy, vol. 123, pp. 1-12, Jun. 2014. [Baidu Scholar] 

8

W. Hu, C. Su, Z. Chen et al., “Optimal operation of plug-in electric vehicles in power systems with high wind power penetrations,” IEEE Transactions on Sustainable Energy, vol. 4, no. 3, pp. 577-585, Jul. 2013. [Baidu Scholar] 

9

C. Jin, T. Jian, and P Ghosh, “Optimizing electric vehicle charging: a customer’s perspective,” IEEE Transactions on Vehicular Technology, vol. 62, no. 7, pp. 2919-2927, Sept. 2013. [Baidu Scholar] 

10

A. Ravey, R. Roche, B. Blunier et al., “Combined optimal sizing and energy management of hybrid electric vehicles,” in Proceedings of 2012 IEEE Transportation Electrification Conference and Expo (ITEC), Dearborn, USA, Jun. 2012, pp. 1-6. [Baidu Scholar] 

11

D. Cao, W. Hu, J. Zhao et al., “Reinforcement learning and its applications in modern power and energy systems: a review,” Journal of Modern Power Systems and Clean Energy, vol. 8, no. 6, pp. 1029-1042, Nov. 2020. [Baidu Scholar] 

12

H. He, C. Wang, H. Jia et al., “An intelligent braking system composed single-pedal and multi-objective optimization neural network braking control strategies for electric vehicle,” Applied Energy, vol. 259, pp. 114-172, Feb. 2020. [Baidu Scholar] 

13

H. Zhou, Y. Zhou, J. Hu et al., “LSTM-based energy management for electric vehicle charging in commercial-building prosumers,” Journal of Modern Power Systems and Clean Energy, vol. 9, no. 5, pp. 1205-1216, Sept. 2021. [Baidu Scholar] 

14

Y. Wu, H. Tan, J. Peng et al., “Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus,” Applied Energy, vol. 247, pp. 454-466, Aug. 2019. [Baidu Scholar] 

15

Z. Wan, H. Li, H. He et al., “Model-free real-time EV charging scheduling based on deep reinforcement learning,” IEEE Transactions on Smart Grid, vol. 10, no. 5, pp. 5246-5257, Sept. 2019. [Baidu Scholar] 

16

A. Chis, J. Lunden, and V. Koivunen, “Reinforcement learning-based plug-in electric vehicle charging with forecasted price,” IEEE Transactions on Vehicular Technology, vol. 66, no. 5, pp. 3674-3684, May 2017. [Baidu Scholar] 

17

Z. Chen, C. Mi, J. Xu et al., “Energy management for a power-split plug-in hybrid electric vehicle based on dynamic programming and neural networks,” IEEE Transactions on Vehicular Technology, vol. 63, no. 4, pp. 1567-1580, May 2014. [Baidu Scholar] 

18

J. Moreno, M. E. Ortuzar, and J. W. Dixon, “Energy-management system for a hybrid electric vehicle using ultracapacitors and neural networks,” IEEE Transactions on Industrial Electronics, vol. 53, no. 2, pp. 614-623, May 2006. [Baidu Scholar] 

19

M. R. Shaarbaf and M. Ghayeni, “Identification of the best charging time of electric vehicles in fast charging stations connected to smart grid based on Q-learning,” in Proceedings of 2018 Electrical Power Distribution Conference (EPDC), Tehran, Iran, May 2018, pp. 78-83. [Baidu Scholar] 

20

R. S. Sutton and A. G. Barto, Introduction to Reinforcement Learning. Cambridge: MIT Press, 1998. [Baidu Scholar] 

21

S. Vandael, B. Claessens, D. Ernst et al., “Reinforcement learning of heuristic EV fleet charging in a day-ahead electricity market,” IEEE Transactions on Smart Grid, vol. 6, no. 4, pp. 1795-1805, Jul. 2015. [Baidu Scholar] 

22

D. Stoyan and L. Redouane, “Reinforcement learning based algorithm for the maximization of EV charging station revenue,” in Proceedings of 2014 International Conference on Mathematics and Computers in Sciences and in Industry, Varna, Bulgaria, Sept. 2014, pp. 235-239. [Baidu Scholar] 

23

D. O’Neill, M. Levorato, A. Goldsmith et al., “Residential demand response using reinforcement learning,” in Proceedings of IEEE SmartGrid Communications 2010, Gaithersburg, USA, Oct. 2010, pp. 409-414. [Baidu Scholar] 

24

D. Osmankovic and S. Konjicija, “Implementation of Q-Learning algorithm for solving maze problem,” in Proceedings of the 34th International Convention MIPRO, Opatija, Croatia, May 2011, pp. 1619-1622. [Baidu Scholar] 

25

V. Mnih, K. Kavukcuoglu, D. Silver et al. (2013, Dec.). Playing atari with deep reinforcement learning. [Online]. Available: https://arxiv.org/abs/1312.5602 [Baidu Scholar] 

26

V. Mnih, K. Kavukcuoglu, D. Silver et al., “Human-level control through deep reinforcement learning,” Nature, vol. 518, pp. 529-533, Feb. 2015. [Baidu Scholar] 

27

D. Silver, A. Huang, C. J. Maddison et al., “Mastering the game of Go with deep neural networks and tree search,” Nature, vol. 529, pp. 484-489, Jan. 2016. [Baidu Scholar] 

28

E. Oh and H. Wang, “Reinforcement-learning-based energy storage system operation strategies to manage wind power forecast uncertainty,” IEEE Access, vol. 8, pp. 20965-20976, Jan. 2020. [Baidu Scholar] 

29

C. Chen, M. Cui, F. Li et al., “Model-free emergency frequency control based on reinforcement learning,” IEEE Transactions on Industrial Informatics, vol. 17, no. 4, pp. 2336-2346, Apr. 2021. [Baidu Scholar] 

30

M. Gheisarnejad, H. Farsizadeh, and M. H. Khooban, “A novel non-linear deep reinforcement learning controller for DC/DC power buck converters,” IEEE Transactions on Industrial Electronics. doi: 10.1109/TIE.2020.3005071 [Baidu Scholar] 

31

Z. Wan, H. Li, H. He et al., “A data-driven approach for real-time residential EV charging management,” in Proceedings of 2018 IEEE PES General Meeting (PESGM’2018), Portland, USA, Jul. 2018, pp. 1-5. [Baidu Scholar] 

32

T. P. Lillicrap, J. J. Hunt, A. Pritzel et al., “Continuous control with deep reinforcement learning,” in Proceedings of ICLR 2016: International Conference on Learning Representations 2016, San Juan, Puerto Rico, May 2016, p. 6. [Baidu Scholar] 

33

V. D. W. Jos and J. Lasenby. (2018, Jan.). The unreasonable effectiveness of the forget gate. [Online]. Available: https://arxiv.org/abs/1804.04849 [Baidu Scholar] 

34

B. Richard, Dynamic Programming. New York: Dover Publications, 1957. [Baidu Scholar] 

35

K. Cho, B. Merrienboer, C. Gulcehre et al., “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, Oct. 2014, pp. 1724-1734. [Baidu Scholar] 

36

W. Zaremba, I. Sutskever, and O. Vinyals. (2014, Jun.). Recurrent neural network regularization. [Online]. Available: https://arxiv.org/abs/1409.2329 [Baidu Scholar] 

37

Z. C. Lipton, J. Berkowitz, and C. Elcan. (2015, Jan.). A critical review of recurrent neural networks for sequence learning. [Online].Available: https://arxiv.org/abs/1506.00019 [Baidu Scholar] 

38

S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, pp. 1735-1780, Sept. 1997. [Baidu Scholar] 

39

A. Graves, A. Mohamed, and G. Hinton, “Speech recognition with deep recurrent neural networks,” in Proceedings of 2013 IEEE International Conference on Acoustics Speech and Signal Processing, Vancouver, Canada, May 2013, pp. 6645-6649. [Baidu Scholar] 

40

F. A. Gers, N. N. Schraudolph, and J. A. Schmidhuber, “Learning precise timing with LSTM recurrent networks,” Journal of Machine Learning Research, vol. 3, pp. 115-143, Mar. 2003. [Baidu Scholar] 

41

F. A. Gers, J. A. Schmidhuber, and F. A. Cummins, “Learning to forget: continual prediction with LSTM,” Neural Computation, vol. 12, pp. 2451-2471, Dec. 2000. [Baidu Scholar] 

42

X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier neural networks,” in Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics (AISTATS-11), Ft. Lauderdale, USA, Nov. 2011, pp. 315-323. [Baidu Scholar] 

43

PJM. (2017, Mar.). Zone COMED. [Online]. Available: https://www.engieresources.com/ [Baidu Scholar] 

44

G. Lai, W. C. Chang, Y. Yang et al., “Modeling long- and short-term temporal patterns with deep neural networks,” in Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, USA, Jun. 2018, pp. 95-104. [Baidu Scholar] 

45

R. Dubey, S. R. Samantaray, and B. K. Panigrahi, “An spatiotemporal information system based wide-area protection fault identification scheme,” International Journal of Electrical Power & Energy Systems, vol. 89, pp. 136-145, Dec. 2017. [Baidu Scholar] 

46

M. Cui, J. Wang, and B. Chen, “Flexible machine learning-based cyberattack detection using spatiotemporal patterns for distribution systems,” IEEE Transactions on Smart Grid, vol. 11, no. 2, pp. 1805-1808, Mar. 2020. [Baidu Scholar] 

47

S. Sun, Q. Yang, and W. Yan. “Optimal temporal-spatial PEV charging scheduling in active power distribution networks,” Protection and Control of Modern Power Systems, vol. 2, no. 1, pp. 1-10, Jan. 2017. [Baidu Scholar] 

48

S. R. Etesami, W. Saad, N. B. Mandayam et al., “Smart routing of electric vehicles for load balancing in smart grids,” Automatica, vol. 120, p. 109148, Oct. 2020. [Baidu Scholar] 

49

T. Ding, Z. Zeng, J. Bai et al., “Optimal electric vehicle charging strategy with Markov decision process and reinforcement learning technique,” IEEE Transactions on Industry Applications, vol. 56, no. 5, pp. 5811-5823, May 2020. [Baidu Scholar]