Journal of Modern Power Systems and Clean Energy

ISSN 2196-5625 CN 32-1884/TK

网刊加载中。。。

使用Chrome浏览器效果最佳,继续浏览,你可能不会看到最佳的展示效果,

确定继续浏览么?

复制成功,请在其他浏览器进行阅读

Doubly-fed Deep Learning Method for Bad Data Identification in Linear State Estimation  PDF

  • Yingzhong Gu (Senior Member, IEEE)
  • Zhe Yu (Senior Member, IEEE)
  • Ruisheng Diao (Senior Member, IEEE)
  • Di Shi (Senior Member, IEEE)
GEIRI North America, San Jose, CA 95134, USA

Updated:2020-11-24

DOI:10.35833/MPCE.2020.000533

  • Full Text
  • Figs & Tabs
  • References
  • Authors
  • About
OUTLINE

Abstract

With more data-driven applications introduced in wide-area monitoring systems (WAMS), data quality of phasor measurement units (PMUs) becomes one of the fundamental requirements for ensuring reliable WAMS applications. This paper proposes a doubly-fed deep learning method for bad data identification in linear state estimation, which can: ① identify bad data under both steady states and contingencies; ② achieve higher accuracy than conventional pre-filtering approaches; ③ reduce iteration burden for linear state estimation; ④ efficiently identify bad data in a parallelizable scheme. The proposed method consists of four key steps: ① preprocessing filter; ② online training of short-term deep neural network; ③ offline training of long-term deep neural network; ④ a decision merger. Through delicate design and comprehensive training, the proposed method can effectively differentiate the bad data from event data without relying on real-time topology information. An IEEE 39-bus system simulated by DSATools TSAT and a provincial electric power system with real PMU data collected are used to verify the proposed method. Multiple test scenarios are applied, which include steady states, three-phase-to-ground faults with (un)successful auto-reclosing, low-frequency oscillation, and low-frequency oscillation with simultaneous three-phase-to-ground faults. The proposed method demonstrates satisfactory performance during both the training session and the testing session.

I. Introduction

NOWADAYS, as artificial intelligence (AI) and big data technologies keep advancing and evolving, more and more data-driven applications are adopted in wide-area monitoring systems (WAMS) [

1]. Operation data with massive volume and velocity keep flowing into WAMS to support applications, e.g., 25-60 samples per second, which requires extensive data-parallel processing capability [2]. Among those requirements, data quality is one of the fundamental demands affecting the quality of decision making and accuracy of system monitoring [3]. In traditional WAMS, linear state estimation (LSE) can be used, by leveraging the linear relationship between the system states (i.e., voltage phasors) and the measurements (i.e., voltage and current phasors) to efficiently calculate the system states and perform bad data identifications [4]. The majority of commercial LSE applications are based on weighted least square (WLS) method, which can only identify and remove one bad measurement per iteration. Otherwise, legit measurements can be treated as bad ones and therefore get removed falsely [5], [6]. This limitation forms a significant bottleneck affecting the self-healing capability of data quality in modern WAMS [7]. To fill this technology gap, this paper presents an innovative method by leveraging deep neural networks to improve bad data processing of phasor measurement unit (PMU) data for LSE used in a control center.

Many recent research works are reported for bad data identification in LSE, which can be generally divided into two categories: state estimation approaches and data-driven approaches [

8]. According to the order of data processing module in the decision-making flow, they can also be classified into the pre-estimation and the post-estimation filtering [9]. Among various state estimation techniques, WLS approach, weighted least absolute value (WLAV) and Bayesian estimation are the popular ones adopted and implemented in WAMS [10].

WLS solves an optimization problem to find the estimated states with the least weighted square error (L2 norm) of the measurement residuals [

11]. Then, bad data can be identified based on the residual between actual measurements and estimated measurements [12]. As an extension of WLS, [11] proposed a distributed WLS state estimation method that integrated weight update from multi-area measurements to enhance the capability and accuracy of handling bad data detection. Reference [12] proposed a substation-level bad data detection algorithm by parsing online state of circuit breakers and disconnectors, which can detect bad data from failing current transformers. Reference [6] presented an alternative WLS largest normalized residual (LNR) algorithm that classifies the suspicious measurements into non-interactive groups and then performs simultaneous identification of multiple bad data.

As an alternative to WLS, WLAV solves an optimization problem to find the estimated states that minimize the weighted absolute value (L1 norm) of the measurement residuals [

13]. Extended from WLAV, the following valuable research works have been conducted. Reference [14] investigated the sparse recovery models for bad data detection and state estimation in power networks where the L1-relaxation model, the multi-stage convex relaxation model and the WLAV were considered. Reference [13] presented a least absolute value based linear state estimator that is computationally efficient and statistically robust.

Different from WLS and WLAV, Bayesian estimation is designed to find the value of states with the highest likelihood [

15]. Reference [15] proposed a Bayesian state estimation for unobservable distribution systems via deep learning, which demonstrates the robustness against modeling and estimation errors and the presence of bad and missing data. Reference [16] developed a Bayesian-based harmonic state estimation in distribution systems to calculate states and identify bad data.

Kalman filter is another popular state estimation approach, especially for power system dynamic state estimation [

17], [18]. Reference [17] developed an extended Kalman filter (EKF) technique for dynamic state estimation of a synchronous machine using PMU data. Reference [18] developed a robust iterated EKF based approach for estimating power system state dynamics subject to disturbances.

In general, the state estimation based approach can efficiently calculate system states. With accurate and timely updated topology information, e.g., contingencies, breaker status, the identification accuracy of bad data can be quite high. However, the state estimation approaches also suffer from drawbacks in handling bad data [

19]. First, these approaches cannot process bad data in critical measurements, critical measurement pair or homogeneously critical measurement groups due to lack of redundancy [20], [21]. Second, the computation time of bad data identification grows linearly to the number of bad data presented and grows exponentially to the system size, or more specifically, the number of measurements and the number of system states [6]. The bad data processing capability for large systems is restrained. Third, the quality of decision making of those approaches relies heavily on accurate and timely updated system topology. Inaccurate or delayed topology information can cause unreliable results in state estimation and bad data identification [19]. Therefore, state estimation based approach alone is not sufficient for online bad data identification in modern WAMS.

The other category of approach is the data-driven approach. Reference [

9] proposed a Kalman filter based pre-estimation approach for bad data identification that can detect abrupt changes among consecutive measurements. Reference [22] proposed an online data-driven algorithm to identify low-quality synchronphasor measurements through a density-based local outlier factor (LOF) analysis. Reference [23] proposed a data-driven PMU bad data detection algorithm based on spectral clustering using single PMU data. Reference [24] proposed a feature-based method originating from simple logical method based on observed patterns. Reference [25] presented a wavelet transformation based approach aiming to alleviate the efforts of system operators. Reference [26] proposed a matrix recovering technique that can be utilized to identify and recover bad data via recognizing the low-rank feature of synchrophasors from adjacent channels. Reference [27] presented a low-pass filter for removing spikes in the measurements. Model-based approaches for bad data identification were also proposed that can improve the accuracy of bad data identification [20], [28], [29]. The data-driven approaches generally have a more efficient performance on processing massive volume of data and are usually positioned as a preprocessing filter for an LSE or other WAMS applications. However, the accuracy of the bad data identification via such approaches usually compromises when events or contingencies happen. It is difficult for such approaches to differentiate the bad data from event data accurately.

This paper proposes a double-fed deep learning method for bad data identification that can accurately label the bad data under both normal operation conditions and various types of events. The proposed method has improved robustness against different kinds of system topological changes. The proposed method can handle critical measurements as long as properly labelled data samples are available. These labelled data samples can be generated by leveraging an offline three-phase LSE, dynamic state estimation on historical PMU or supervisory control and data acquisition (SCADA) data and numerical simulations.

The key contributions of this paper are threefold: ① a double-fed deep learning method for bad data processing is proposed to efficiently label bad data for large systems in parallelizable scheme; ② the proposed method can accurately identify bad data under normal operation conditions and various types of events and contingencies; ③ the proposed method has improved robustness against different kinds of system topological changes.

The remainder of the paper is organized as follows. Section II briefly discusses the formulations of LSE. Section III introduces the proposed double-fed deep learning method for bad data preprocessing. Section IV presents the numerical experiments of the proposed method under different scenarios using practical system PMU data. Section V provides the concluding remarks and discusses potential future works.

II. LSE

Since the bad data processing approach is designed for LSE, this section briefly presents the concept of LSE. LSE calculates system states based on phasor measurements by leveraging the linear relationship between the voltage and current phasors [

4]. The advantages of LSE includes: ① non-iterative solution for state estimation; ② efficient computation performance compared with conventional state estimation; ③convergence guarantee compared with Newton method based state estimation [30]. Popular approaches for state estimation include WLS [7], [10], [26], WLAV [13], [14], Bayesian estimation [15], [16] and Kalman filter estimation [17], [18]. In industry applications, WLS is most widely adopted. Therefore, this section takes the WLS as an example to illustrate the LSE.

Assuming a system with n observable nodes, mV voltage measurements and mI branch current measurements, the total number of measurements can be denoted as m, which is the summation of mV and mI. The system state vector x˙n×1 includes the voltage phasors of all the observable nodes. The measurement vector z˙m×1 includes the voltage and current phasor measurements of the terminals where PMUs are installed. The measurement model of PMU data can be derived from Ohm’s law as

V˙=Ax˙I˙=Y˙x˙ (1)

where AmV×n is the identity matrix which describes the one-to-one mapping relationship between the state vector x˙ and voltage phasor measurement vector V˙, for those buses where PMU measurement is available; and Y˙mI×n is the branch admittance matrix which is derived from Kirchhoff’s voltage law (KVL) to describe the relationship between the branch current phasor I˙ and the two voltage phasors at the two terminals. If the voltage phasor of node j is the ith component in the measurement vector of voltage phasors, Ai,j=1; otherwise Ai,j=0, where Ai,j is the element of A on the ith row and jth column. By combining the voltage and current measurements into one formulation, the measurement model of PMU data can be denoted by the complex matrix H˙ in (2).

z˙=V˙I˙=AY˙x˙=H˙x˙ (2)

The model in (2) is not numerically linear yet, because its components x and z are complex numbers. It needs to be further decomposed into the rectangular form in (3). The corresponding measurement model becomes (4).

x=Re(x˙)Im(x˙)z=Re(z˙)Im(z˙) (3)
z=HR-HIHIHRx=Hx (4)

where H is the matrix which represents the measurement model for LSE in rectangular form; and HR and HI are the real and imaginary parts of the H, respectively [

31].

Based on the formulation in (4), the solution of x, namely the vector of estimated system states, can be obtained by (5).

x̂=(HTW-1H)-1HTW-1z (5)

where Wm×m is a diagonal matrix, whose diagonal components are the weights of the measurements.

The LSE form (3)-(5) is the positive-sequence LSE. Although this is widely used in practical power industry due to its simplicity and efficiency, it is more useful for large-scale system during steady state. For the purpose of bad data identification, an alternative form of the three-phase representation is preferred because it can correlate and crosscheck all three phases [

32].

By extending the positive-sequence voltage and current phasors to three-phase voltage and current forms, (1) can be denoted as

V˙abc=V˙aV˙bV˙cT=Aabcx˙abc (6)
I˙abc=I˙aI˙bI˙cT=Y˙abcx˙abc (7)

where V˙a, V˙b, V˙c are the three-phase voltage phasors; I˙a, I˙b, I˙c are the three-phase current phasors; Aabc3mV×(3n) is the augmented identity matrix; Y˙abc3mI×(3n) is the augmented branch admittance matrix; and x˙abc is the three-phase system state vector.

To further illustrate (7), one of the branches can be picked and (7) can be represented as

I˙ijaI˙ijbI˙ijcI˙jiaI˙jibI˙jic=y˙ijabc+y˙i0abc-y˙ijabc-y˙ijabcy˙ijabc+y˙j0abcV˙iaV˙ibV˙icV˙jaV˙jbV˙jc (8)

where I˙ijd is the from-end current phasor of phase d, d{a,b,c}; I˙jid is the to-end current phasor of phase d; y˙ijabc is the branch admittance parameters; y˙i0abc is the front-end branch to ground susceptance; y˙j0abc is the to-end branch to ground susceptance; V˙id is the front-end voltage phasor at phase d; and V˙jd is the to-end voltage phasor at phase d.

Equation (2) can be extended to

z˙abc=V˙abcI˙abc=AabcY˙abcx˙abc=H˙abcx˙abc (9)

where z˙abc is the phasor vector of three-phase measurements; and H˙abc is the measurement matrix for the three-phase LSE.

Equations (6) and (7) can be converted to the rectangular form (10), and (9) can be written as (11).

xabc=Re(x˙abc)Im(x˙abc)zabc=Re(z˙abc)Im(z˙abc) (10)
zabc=HR,abc-HI,abcHI,abcHR,abcxabc=Habcxabc (11)

where xabc is the scalar vector including both real and imaginary parts of the state variables; zabc is the scalar vector including both real and imaginary parts of the measurement variables; and HR,abc and HI,abc are real and imaginary parts of H˙abc, respectively.

Based on the formulation in (11), the solution of xabc can be obtained by

x̂abc=(HabcTWabc-1Habc)-1HabcTWabc-1zabc (12)

where Wabc(3m)×(3m) is a diagonal matrix, whose diagonal components are the weights of the measurements at three phases.

The three-phase gain matrix can be defined as

Gabc=HabcTWabc-1Habc (13)

The three-phase hat matrix can be defined as

Kabc=HabcGabc-1HabcTWabc (14)

Then, the three-phase sensitivity matrix can be calculated as (15), where I is the identitiy matrix.

Sabc=I-Kabc (15)

The residuals of the measurement can be calculated by the difference between the actual and the estimated measurements in (16).

rabc=zabc-x̂abcHabc (16)

The covariance matrix of the measurement residual can be calculated as

Ω=SabcWabc-1 (17)

The normalized residual for each measurement can be calculated as

rN,i=|ri|Ωii    i[1,m] (18)

where ri is the residual of each measurement; and Ωii is the diagonal element of the covariance matrix Ω.

In this paper, the measurement value whose normalized residual is above 3.0 is considered as bad data or abnormal data. In that case, the probability of that measurement is estimated to be less than 0.3% given that the measurement error follows a normal distribution, which is a widely adopted standard in the industry [

30]. The three-phase LSE is more effective in identifying bad data because bad data occurrrence violating three-phase balancing can be identified [32]. However, due to the problem dimension being at least nine times larger than the positive-sequence state estimation, its online applicability and feasibility in practical systems, especially those large regional systems, are compromised [33]. In this paper, we use the three-phase LSE to generate bad data labels for practical PMU data, which can be employed to train the AI agent using the proposed method. Then online bad data identifications can be performed by a well-trained agent in a distributed manner for large-scale systems in real time.

III. Doubly-fed Deep Learning Method for Bad Data Identification

The proposed doubly-fed deep learning method for bad data identification consists of two feedback-pipelines: short-term deep neural network and long-term deep neural network. The short-term deep neural network is designed to learn and capture short-term data patterns under variable situations. Online training is performed for the short-term deep neural network to keep updated to the current system operation point. The long-term deep neural network is designed to learn and capture historical patterns and theoretically presumptive patterns such as contingencies or various events generated through transient analysis simulations.

Figure 1 shows the architecture of the proposed method. The green arrow bar represents the incoming PMU data stream. Two moving windows take consecutive data frames and feed them into the corresponding pre-filter. For short-term deep neural network pipeline, the pre-filter S processes the data frames from moving window S and feeds them into the deep neural network S. For long-term deep neural network pipeline, the pre-filter L processes the data frames from the moving window L and feeds them into the deep neural network L. The moving window L has a length twice longer than the moving window S because it needs more data to diagnose whether the current pattern is similar to any of the historical patterns. Both deep neural networks S and L generate the probability of the current data samples being bad data for every channel. The decision functions L and S convert the probability into the corresponding decision scores. The decision merger generates a decision label for whether the current data sample is a bad data for each channel based on the decision scores from both the short-term and long-term deep neural network pipelines. The decision labels are passed to the linear state estimator, which can remove the bad data before the first-round calculations. The estimated states and measurements are written into the LSE stream, which feeds back to the pre-filters L and S to remove the bad data that may affect the scores for other data samples. The LSE stream also feeds them into the deep neural network S to perform periodically online training.

Fig. 1 Flow chart of deep learning method for bad data identification.

The data structure of PMU measurements feeding into the moving windows S and L is illustrated in Fig. 2, where Va, Vb, Vc, θva, θvb, θvc are the magnitudes and phase angles of the three-phase voltage phasors, respectively; Ia, Ib, Ic, θIa, θIb, θIc are the magnitudes and phase angles of the three-phase current phasors, respectively; and f is the frequency. PMUs are grouped into clusters for the double-fed deep neural networks. There are two principles for the PMU clustering.

Fig. 2 PMU data structure and moving window.

1) Correlation principle: the grouped PMU measurements should have a strong correlation to each other governed by the physical laws of nature.

2) Independence principle: the errors and noises of the grouped PMU measurements should have independent probability distributions.

Good examples of clustering include independent PMU measurements at the same substation but for different transmission lines, independent PMU measurements at two terminals of a branch, independent measurements of parallel circuits, and independent measurements at different windings of a transformer. Bad examples of clustering can be irrelevant measurements of elements with weak correlation, correlated measurements sharing the same potential transformers (PTs), current transformers (CTs), or communication channels.

The number of clustered PMU k is suggested to be between one and three. If only one PMU is used, the deep neural network can only learn the cross-correlation between voltages and currents, between phase angles and magnitudes, and among the three phases. With independent PMU measurement incorporated into the deep neural networks, data can be validated across different circuits, PTs, CTs, and terminals, which significantly improve the accuracy of bad data identification. However, if there are too many PMUs clustered in the same deep neural network, the computation performance may decline, and the neural network may not work properly if one of the PMUs is off-line. After the clusters are formed, all 13 measurement quantities in Fig. 2 are integrated into a vector to be fed into the deep neural network as an input vector. The pre-filter S processes time-series data in different PMU channels independently in a parallel manner, while the deep neural network takes pre-filters’ outputs clustered altogether.

The moving window can be denoted as

W=wt|tt0+tq-L,...,t0,...,t0+tq (19)

where W is the set of the data in the moving window; t is the time stamp; t0 is the target time for LSE; tq is the lead time steps (suggested range is 60-100 ms); and L is the length of the window. The lead time can be set as zero if system operators intend to calculate the LSE solution immediately once the PMU measurement data is available. However, it has been found through many numerical experiments that a non-zero lead time brings the advantage of identifying temporary bad data spikes that can effectively improve the resulting quality of LSE.

The pre-filter can be applied via

Q(p)=inf{wt: pF(wt): wtW} (20)
Pf(wt)=2g(2(wt-Q(p0)))-1g(wt)=ewt1+ewt (21)

where F(·) is the cumulative probability distribution function of the data set W; p is the threhold probability; p0 is the default threhld probability; inf(·) is the infimum of the set or the greatest lower bound of the set; Q(·) is the function which calculates the modified current inputs; g(·) is the logistic sigmoid function; and Pf(·) is the function which gives the inputs for the deep neural network. This proposed moving window can leverage the temporal correlation of the data sets to capture temporarily false spikes and mitigate negative impacts of extreme data points on other non-extreme data points.

Based on the numerical experiments, the pre-filter process described in (20) and (21) is essential for the successful functioning of the proposed method. It mitigates the dependency of deep neural network training results on the power system operation conditions and develops a heterogeneous sensitivity toward different levels of measurement errors. The output of Pf(·) is normalized as the interval of (-1,1).

The outputs of the pre-filter feeding into the deep neural network are depicted in Fig. 3. Assuming that k PMU measurements are grouped in the cluster, the dimension of the inputs to this deep neural network is 13k, which includes the three-phase voltage magnitudes and angles, three-phase current magnitudes and angles, and the frequency. The number of hidden layers is suggested to be 4, the activation function of the first layer is tanh, the activation function of the rest hidden layers is relu, the activation function of the output layer is softmax, the number of units for each hidden layer is 10k, and the number of units for the output layer is 13k. For each hidden layer, there is a neuron with no input which represents the bias term for that layer. The output of the deep neural network is one-to-one mapping to the input of the network, which indicates the probability of each input being bad data.

Fig. 3 Deep neural network for bad data identification.

L=-1Ni=1Nyiln(p(yi))+(1-yi)ln(1-p(yi)) (22)

where yi is the ith output of the deep neural network.

The binary cross-entropy function (22) is used as the loss function. Based on the numerical experiments, the cross-entropy function has much better performance in both training and testing phases compared with other popular loss functions such as L2 loss, because the output decision is discrete and the convexity of the binary cross-entropy function makes it easier to find the global optimum [

34]. The Adam optimizer is used during the training process, where the learning rate is set to be 0.01.

In the designed architecture presented in Fig. 1, two deep neural networks are used. The short-term neural network generates a probability output using PMU data from the moving window S. This neural network is re-trained online every 1 min using the last 2-minute PMU data with LSE labels.

The probability decision function is provided by

D(yi)=eβ(yi-0.5)1+eβ(yi-0.5) (23)

where β is the squashing coefficient which is suggested to be 10. The decision function makes the probability converge more quickly to 1 (or 0) when the result goes higher (or lower) than the threshold of 0.5, which is to seek a reasonable balance between exploration and exploitation.

The decision merger combines the two probability decisions D(ys) and D(yl) from the short-term and long-term deep neural network pipelines, as described by

M(yi)=(D(ys)D(yl))12 (24)

Different weights can be assigned to demonstrate certain preferences in decision making toward a long-term perspective or short-term perspective. Since LSE has its own bad data identification, it is preferred to be conservative for the neural network to avoid false alarms because any bad data identified by the neural network will be removed directly. For those unidentified bad data, the LSE will process them. Although the proposed method can process most of bad data, some remaining ones (due to small deviations or scenarios not included in the training set) still need to be identified in LSE and are subject to the limitations of LSE.

IV. Numerical Experiments

In this paper, the proposed method has been applied to the IEEE 39-bus system using TSAT simulation and one of the largest provincial power systems in China–the Jiangsu power system using the practical PMU data.

A. IEEE 39-bus System

The IEEE 39-bus system is a 10-machine New England power system. In the system, generator 1 represents an aggregation of generators. The detailed system parameters are available at [

35]. In the IEEE 39-bus system, generators are in a fourth-order model, while the constant impedance load model is adopted. The system has a single oscillatory mode with natural frequency of f0=1.3217 Hz. A sinusoidal signal Δref=ksin(2ft) with k=0.6 is added to the reference signal of excitation systems. Generator 1 acts as the source of the forced oscillation. The oscillation disturbance is added to the system at t=0 s. The low-frequency oscillation event and three-phase-to-ground fault are simulated by DSATools TSAT v18.1.37.

Figure 4 illustrates the current magnitudes of branch 1-2 during a low-frequency oscillation that occurs at generator 1, where the red dot represents the bad data label and the blue bar represents the bad data identification probability of the proposed method (Fig. 5, Fig. 7-11 follow the same legends). Once the blue bar is above 50%, the data sample is labelled as bad data by the proposed method. These samples won’t be taken by LSE as legit measurements in the subsequent computation. The goal of the proposed method is to generate as many blue bars (>50%) to match the red dots at 1.0 as possible and generate as few blue bars whenever the corresponding red dots are at 0 as possible. As shown in Fig. 4, the proposed method successfully avoids labeling oscillation data samples as bad data during the entire oscillation period. Meanwhile, almost all the major bad data at different phases during different periods are successfully labelled, although some identification probabilities are not very large (e.g., phase c at 3 s) and some false alarms are observed (e.g., phase c at 0.6 s). For preprocessing filters, it is acceptable not to label all the bad data because the LSE can process the ones falling through the crack. It is also encouraged not to falsely identify too many good data as bad data because the LSE is unable to correct those false alarms, especially if any of them is critical measurement or in one of the critical pairs.

Fig. 4 Bad data identification probability for three-phase current magnitudes of IEEE 39-bus system during low-frequency oscillation event. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 5 Bad data identification probability for three-phase voltage magnitudes during low-frequency oscillation with a simultaneous three-phase-to-ground fault. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 6 Node-breaker model of 500 kV transmission lines LM #5321 and GL #5322.

Fig. 7 Bad data identification probability for three-phase voltage magnitudes during steady state. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 8 Bad data identification probability for three-phase current magnitudes during non-permanent three-phase-to-ground fault. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 9 Bad data identification probability for three-phase current phase angles during permanent three-phase-to-ground fault with auto-reclosing failure. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 10 Bad data identification probability for three-phase current magnitudes during low-frequency oscillation event. (a) Phase a. (b) Phase b. (c) Phase c.

Fig. 11 Bad data identification probability for three-phase current phase angles during low-frequency oscillation event. (a) Phase a. (b) Phase b. (c) Phase c.

Figure 5 illustrates three-phase voltage magnitudes at bus 1 during a low-frequency oscillation at generator 1 plus a temporary three-phase-to-ground fault at branch 2-3 at 0.1 s. After five cycles, the breakers at both terminals of branch 2-3 are opened by the relay, and then the temporary fault is cleared. After 50 cycles, the transmission line is successfully reclosed by the auto-recloser. As shown in Fig. 5, the proposed method successfully avoids labeling the event data as bad data during the oscillation and topology changes. Meanwhile, almost all the major bad data at different phases during different periods are successfully labelled.

B. Jiangsu Power System

The Jiangsu power system has 731 substations and 349 power plants, where 244 of the substations have been equipped with 1138 PMUs. Each PMU measures three-phase voltage phasors, current phasors, and frequency. There are 30365 nodes, 8557 breakers, 22084 disconnectors, 2335 buses, 2393 transmission lines and 1796 transformers in the node-breaker model of this power system. PMU measurements collected in the system in 2019 are used and have a reporting rate of 25 Hz. Eighty percent of the data sets are used for training for the long-term deep neural network, while the remaining 20% are used for testing. Among the testing data sets, half of the data are used for development phase and the other half are used for testing only. The online training for the short-term deep neural network uses the testing data. The true value of practical PMU measurements is unknown. Therefore, an LSE with bad data identification is used to generate the labels for both the training and testing data. For the proposed method, after training, the average online computation time is 19.3995 μs per sample, which is less than 0.1% of the average computation time of LSE.

To illustrate the effectiveness of the proposed method, a 500 kV double-circuit transmission line is presented in Fig. 6.

In the setup of numerical experiment, all the contingencies occur on GL #5322, while LM #5321 is the monitored line for bad data identification and verification. The following scenarios are considered in the numerical experiments.

1) No contingency occurs, and the system is running in steady state with occasional bad data injections.

2) Temporary three-phase-to-ground fault occurs at GL #5322, the relay clears the fault, and the auto-recloser recloses the line successfully.

3) Permanent three-phase-to-ground fault occurs on GL #5322, the relay isolates the fault, the auto-recloser fails to clear the fault, and the relay opens the line permanently.

4) Low-frequency oscillation occurs at the WT plant #3.

Figure 7 illustrates the non-contingency scenario.

It is worth noting that the phase angles measured by PMU are the original phase angles that are not the angle difference to the reference bus. The values of these angles vary across time depending on the deviations of the actual system frequency. Besides, the phase angle is a cyclical variable in the interval of (-180°, 180°]. As shown in Fig. 7, the proposed method works pretty well under steady state, and almost all the bad data samples can be labelled correctly.

Figure 8 illustrates the three-phase current magnitudes of LM #5321 when the temporary three-phase-to-ground fault occurs at GL #5322 at 0.1 s. The fault location is 0.1 km from LG plant #2. After five cycles (the base frequency of system is 50 Hz), the breakers at both terminals of GL #5322 are opened by the relay. After that, the temporary fault is cleared, and 50 cycles later, the transmission line is successfully reclosed by the auto-recloser. As is observed in Fig. 8, when the fault occurs, the proposed method successfully avoids labeling the event spikes causing by three-phase-to-ground fault as bad data, which is a common issue faced by many other numerical approaches or topology information-dependent approaches.

In this paper, the fundamental philosophy of the deep neural network that can work here is that every fault in reality has its unique pattern. If the deep neural network can learn these patterns through comprehensive training of historical event data and simulated data, it can distinguish the event data from the bad data. More specifically, for the event that the current magnitude of phase c has a dip of 0.14 p.u. at 2.6 s, it is more likely to be recognized as a bad data instead of faults or other events.

This is because in a practical system, it is very rare to have a fault or an event only causing the change in one phase without any changes in the other two phases. Even a single-phase-to-ground fault affects the magnitudes and phase angles at all three phases. The three-phase patterns can be learned by the deep neural network. In this scenario, the deep neural network does not falsely label any event data as bad data during both the fault and the reclosing periods. Meanwhile, almost all the bad data at different phases during different periods are labelled successfully, although some values are not very large (e.g., phase B at 2.3 s) and some other bad data are not captured due to their small deviations (e.g., phase b at 3.8 s).

Figure 9 illustrates the current phase angles of LM #5321 during the permanent three-phase-to-ground fault that occurs at GL #5322 at 0.1 s.

The fault location is 0.1 km from LG plant #2. After five cycles (the base frequency of system is 50 Hz), the breakers at both terminals of GL #5322 are opened by the relay. The permanent fault fails to be cleared after that. After 50 cycles, the transmission line is reclosed by the auto-recloser, and the fault is activated again. After another 5 cycles, the breaker is triggered by the relay to open the transmission line permanently. As shown in Fig. 9, the proposed method successfully avoids labeling the event data as bad data at the beginning of the fault, at the 1st time of relay operation, at the 1st time of auto-recloser operation, or at the 2nd time of relay operation. Meanwhile, almost all the major bad data at different phases during different periods are labelled successfully, although some values are not very large (e.g., phase a at 4.0 s, phase c at 1.9 s) and some false alarms almost cross the threshold of 0.5 (e.g., phase a at 0.38 s, phase c at 0.2 s). It is also observed that some values have a certain delay in responsiveness (e.g., phase c at 1.1 s).

Figures 10 and 11 present the bad data identification for current magnitudes and current phase angles during a low-frequency oscillation event, respectively. The low-frequency oscillation event occurs at WT plant #3 at 41 s. The location is 82.8 km southeast to LG plant #2. The frequency of the oscillation is 0.66 Hz and the damping factor is 0.99.

As shown in Fig. 10, the deep learning algorithm successfully labels most of the bad data correctly and avoids labeling the event data points caused by the oscillation event as bad data before, after or during the oscillation event.

As shown in Fig. 11, although the impact of the oscillation on current phase angles is not as obvious as the one on current magnitude, most of the bad data are successfully labelled by the proposed method.

Table I illustrates the performance comparison between the proposed method and the benchmark for the training, development and testing data sets. The benchmark is a five-layer (four hidden layers, 53 neurons, the same dimension as the proposed deep neural network) deep neural network without the proposed double-fed framework. Compared with the proposed method, the training accuracy of the benchmark is only 0.8974, and the development and testing accuracies are only 0.7952 and 0.5367, respectively. The precision, recall and F1 scores of the benchmark are only 0.5306, 0.7222, and 0.4194, respectively. As shown Table I, the performance is improved by the proposed method.

Table I Performance Comparison of Proposed Method and Benchmark
MethodTraining accuracyDevelopment accuracyTesting accuracy

Precision

score

Recall scoreF1 score
Proposed 0.9796 0.9798 0.9577 0.9929 0.9061 0.9475
Benchmark 0.8974 0.7952 0.5367 0.5306 0.7222 0.4194

Table II presents the comparison of computation efficiency between the proposed method and the method of typical filter + LSE. The LSE is equipped with typical PMU data pre-filters including three-phase balancing check, voltage and current typical range check, topology crosscheck, etc. It is observed that the pre-filtering time of the proposed method is almost 3 times longer than running these simple pre-filters. In return, the proposed method can help LSE reduce computation time and the number of iterations significantly so that the total computation time is 57.1273% lower. Most of the model enhancements have been discussed in Section III. Based on extensive numerical experiments, it shows that the moving window, PMU clustering, pre-filtering process, decision function, decision merger and the feedback from LSE results to both long-term and short-term deep neural networks are important and need to be carefully configured, which yields an overall well-performing double-fed deep neural network.

Table II Comparison of Computation Efficiency
MethodAverage LSE iterationAverage pre-filter time (ms)Average LSE time (ms)Average total time (ms)
Proposed method 3.6144 6.3518 25.1415 31.4933
Typical filter + LSE 10.3416 2.3251 71.1321 73.4572

V. Conclusion

As large volumes of PMU data feed into modern control centers, it poses more challenges to the data quality management capability in WAMS applications. Many existing approaches have been developed to handle bad data problems, but it is usually challenging to identify bad data during contingencies, events or topological changes. This paper proposes a doubly-fed deep learning method for bad data identification in LSE. It integrates a long-term deep neural network that learns various historical events/contingencies and a short-term deep neural network that keeps learning the prevailing system operation conditions, and makes reasonably good decisions for labeling bad data. The proposed method is verified using IEEE 39-bus system simulated in TSAT and the real PMU measurements from one of the largest provincial power systems in China. Numerical experiment of normal steady-state and four additional contingencies have illustrated that the proposed method can not only identify the bad data correctly but also not falsely label the event data as bad ones.

Future work can be extended to investigate other machine learning models, e.g., long short-term memory (LSTM), convolutional neural network (CNN), graph neural network (GNN), deep reinforcement learning (DRL), on handling the bad data identification to achieve higher precision and recall scores. It is also worth exploring the potential deep learning applications for dynamic LSEs and full-AC state estimations.

References

1

W. Wang, J. Zhao, W. Yu et al., “FNETVision: a WAMS big data knowledge discovery system,” in Proceedings of 2018 IEEE PES General Meeting (PESGM), Portland, USA, Aug. 2018, pp. 1-5. [百度学术

2

P. H. Gadde, M. Biswal, S. Brahma et al., “Efficient compression of PMU data in WAMS,” IEEE Transactions on Smart Grid, vol. 7, no. 5, pp. 2406-2413, Sept. 2016. [百度学术

3

A. Sundararajan, T. Khan, A. Moghadasi et al., “Survey on synchrophasor data quality and cybersecurity challenges, and evaluation of their interdependencies,” Journal of Modern Power Systems and Clean Energy, vol. 7, no. 3, pp. 449-467, May 2019. [百度学术

4

L. Zhang, H. Chen, K. Martin et al., “Successful deployment and operational experience of using linear state estimator in wide-area monitoring and situational awareness projects,” IET Generation, Transmission & Distribution, vol. 11, no. 18, pp. 4476-4483, Dec. 2017. [百度学术

5

K. D. Jones, A. Pal, and J. S. Thorp, “Methodology for performing synchrophasor data conditioning and validation,” IEEE Transactions on Power Systems, vol. 30, no. 3, pp. 1121-1130, May 2015. [百度学术

6

Y. Lin and A. Abur, “A highly efficient bad data identification approach for very large scale power systems,” IEEE Transactions on Power Systems, vol. 33, no. 6, pp. 5979-5989, Nov. 2018. [百度学术

7

L. M. Putranto, R. Hara, H. Kita et al., “Series PMU data-based state estimation technique for WAMS application,” in Proceedings of 2016 IEEE PES General Meeting (PESGM), Boston, USA, Aug. 2016, pp. 1-5. [百度学术

8

A. Monticelli, “Electric power system state estimation,” Proceedings of the IEEE, vol. 88, no. 2, pp. 262-282, Feb. 2000. [百度学术

9

M. Pignati, L. Zanni, S. Sarri et al., “A pre-estimation filtering process of bad data for linear power systems state estimators using PMUs,” in Proceedings of 2014 Power Systems Computation Conference, Wroclaw, Poland, Aug. 2014, pp. 1-8. [百度学术

10

P. A. Pegoraro, A. Angioni, M. Pau et al., “Bayesian approach for distribution system state estimation with non-gaussian uncertainty models,” IEEE Transactions on Instrumentation and Measurement, vol. 66, no. 11, pp. 2957-2966, Nov. 2017. [百度学术

11

J. Kang and D. Choi, “Distributed multi-area WLS state estimation integrating measurements weight update,” IET Generation, Transmission & Distribution, vol. 11, no. 10, pp. 2552-2561, Jul. 2017. [百度学术

12

Y. Wu, Y. Xiao, F. Hohn et al., “Bad data detection using linear WLS and sampled values in digital substations,” IEEE Transactions on Power Delivery, vol. 33, no. 1, pp. 150-157, Feb. 2018. [百度学术

13

C. Xu and A. Abur, “A fast and robust linear state estimator for very large scale interconnected power grids,” IEEE Transactions on Smart Grid, vol. 9, no. 5, pp. 4975-4982, Sept. 2018. [百度学术

14

J. Yang, W. Wu, W. Zheng et al., “ Performance analysis of sparse recovery models for bad data detection and state estimation in electric power networks,” in Proceedings of 2016 IEEE PES General Meeting (PESGM), Boston, USA, Jul. 2016, pp. 1-5. [百度学术

15

K. R. Mestav, J. Luengo-Rozas, and L. Tong, “Bayesian state estimation for unobservable distribution systems via deep learning,” IEEE Transactions on Power Systems, vol. 34, no. 6, pp. 4910-4920, Nov. 2019. [百度学术

16

W. Zhou, O. Ardakanian, H. Zhang et al., “Bayesian learning-based harmonic state estimation in distribution systems with smart meter and DPMU data,” IEEE Transactions on Smart Grid, vol. 11, no. 1, pp. 832-845, Jan. 2020. [百度学术

17

E. Ghahremani and I. Kamwa, “Dynamic state estimation in power system by applying the extended Kalman filter with unknown inputs to phasor measurements,” IEEE Transactions on Power Systems, vol. 26, no. 4, pp. 2556-2566, Nov. 2011. [百度学术

18

J. Zhao, M. Netto, and L. Mili, “A robust iterated extended Kalman filter for power system dynamic state estimation,” IEEE Transactions on Power Systems, vol. 32, no. 4, pp. 3205-3216, Jul. 2017. [百度学术

19

J. Krstulovic, V. Miranda, A. J. A. S. Costa et al., “Towards an auto-associative topology state estimator,” IEEE Transactions on Power Systems, vol. 28, no. 3, pp. 3311-3318, Aug. 2013. [百度学术

20

S. Pal, B. Sikdar, and J. H. Chow, “Classification and detection of pmu data manipulation attacks using transmission line parameters,” IEEE Transactions on Smart Grid, vol. 9, no. 5, pp. 5057-5066, Sept. 2018. [百度学术

21

X. Wang, D. Shi, J. Wang et al., “Online identification and data recovery for PMU data manipulation attack,” IEEE Transactions on Smart Grid, vol. 10, no. 6, pp. 5889-5898, Nov. 2019. [百度学术

22

M. Wu and L. Xie, “Online identification of bad synchrophasor measurements via spatio-temporal correlations,” in Proceedings of 2016 Power Systems Computation Conference (PSCC), Genoa, Italy, Jun. 2016, pp. 1-7. [百度学术

23

Z. Yang, H. Liu, T. Bi et al., “Bad data detection algorithm for PMU based on spectral clustering,” Journal of Modern Power Systems and Clean Energy, vol. 8, no. 3, pp. 473-483, May 2020. [百度学术

24

C. Wan, H. Chen, M. Guo et al., “Wrong data identification and correction for WAMS,” in Proceedings of 2016 IEEE PES Asia-Pacific Power and Energy Engineering Conference (APPEEC), Xi’an, China, Oct. 2016, pp. 1903-1907. [百度学术

25

H. Li, “A method of bad data identification based on wavelet analysis in power system,” in Proceedings of 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), Zhangjiajie, China, May 2012, pp. 146-150. [百度学术

26

M. Liao, D. Shi, Z. Yu et al., “Estimate the lost phasor measurement unit data using alternating direction multipliers method,” in Proceedings of 2018 IEEE/PES Transmission and Distribution Conference and Exposition (T&D), Denver, USA, Apr. 2018, pp. 1-9. [百度学术

27

X. Deng, D. Bian, D. Shi et al., “Impact of low data quality on disturbance triangulation application using high-density PMU measurements,” IEEE Access, vol. 7, pp. 105054-105061, Jul. 2019. [百度学术

28

D. Shi, D. J. Tylavsky, and N. Logic, “An adaptive method for detection and correction of errors in PMU measurements,” IEEE Transactions on Smart Grid, vol. 3, no. 4, Vancouver, Canada, Jul. 2012, pp. 1575-1583. [百度学术

29

X. Wang, D. Shi, Z. Wang et al., “Online calibration of phasor measurement unit using density-based spatial clustering,” IEEE Transactions on Power Delivery, vol. 33, no. 3, pp. 1081-1090, Jun. 2018. [百度学术

30

L. Zhang, A. Bose, A. Jampala et al., “Design, testing, and implementation of a linear state estimator in a real power system,” IEEE Transactions on Smart Grid, vol. 8, no. 4, pp. 1782-1789, Jul. 2017. [百度学术

31

A. G. Phadke and J. S. Thorp, Synchronized Phasor Measurements and Their Applications. Berlin: Springer, 2008. [百度学术

32

K. D. Jones, J. S. Thorp, and R. M. Gardner, “Three-phase linear state estimation using phasor measurements,” in Proceedings of 2013 IEEE PES General Meeting, Vancouver, Canada, Aug. 2013, pp. 1-5. [百度学术

33

T. Yang, H. Sun, and A. Bose, “Transition to a two-level linear state estimator–Part I: architecture,” IEEE Transactions on Power Systems, vol. 26, no. 1, pp. 46-53, Feb. 2011. [百度学术

34

Z. Zhang and M. Sabuncu, “Generalized cross entropy loss for training deep neural networks with noisy labels,” in Advances in Neural Information Processing Systems, New York: Curran Associates, Inc., 2018, pp. 8778-8788. [百度学术

35

T. Athay, R. Podmore, and S. Virmani, “A practical method for the direct analysis of transient stability,” IEEE Transactions on Power Apparatus and Systems, vol. PAS-98, no. 2, pp. 573-584, Mar. 1979. [百度学术