Abstract
Real-time voltage stability assessment (VSA) has long been an extensively research topic. In recent years, rapidly mounting deep learning methods have pushed online VSA to a new height that large amounts of learning algorithms are applied for VSA from the perspective of measurement data. Deep learning methods generally require a large dataset which contains measurements in both secure and insecure states, or even unstable state. However, in practice, the data of insecure or unstable state is very rare, as the power system should be guaranteed to operate far away from voltage collapse. Under this circumstance, this paper proposes an autoencoder based method which merely needs data of secure state to evaluate voltage stability of a power system. The principle of this method is that an autoencoder purely trained by secure data is expected to only create precise reconstruction for secure data, while it fails to rebuild data of insecure states. Thus, the residual of reconstruction is effective in indicating VSA. Besides, to develop a more accurate and robust algorithm, long short-term memory (LSTM) networks combined with fully-connected (FC) layers are used to build the autoencoder, and a moving strategy is introduced to bias the features of testing data toward the secure feature domain. Numerous experiments and comparison with traditional machine learning algorithms demonstrate the effectiveness and high accuracy of the proposed method.
WITH the continuous increase of the penetration of renewable generations and flexible consumers, power systems are more likely to operate near the voltage collapse point (VCP), which makes the voltage stability more critical for the security and economy of modern power systems [
In recent years, with the wide deployment of phasor measurement unit (PMU), a huge amount of high-resolution data has been collected, which prompts many researchers to study data-driven methods for VSA. Data-driven methods analyze the behaviour of power systems from the perspective of measuring data, and they require no prior knowledge of the complex model and parameters of power systems. Therefore, it can circumvent the information loss caused by manual assumptions and simplifications which generally occur when using traditional model-based methods [
Even though most of literature has obtained promising accuracy, one critical drawback is that their performance highly depends on a large training data set, which must contain the data of each stage of the voltage stability deterioration process. Namely, from the view of P-V curve, data set is required to include certain data of each operation point above the VCP. However, in practice, the data of insecure state, i.e., at a point close to the VCP, is very rare, since power systems normally operate in secure state [
Faced with this dilemma, this paper proposes a method which avoids using the data of insecure state, i.e., this method is purely trained by the data of secure state which is easily accessible. The idea behind the proposed method comes from a new perspective that the reconstruction loss of a well-modified autoencoder is effective to indicate the change of data distributions [
The main contributions of our work are summarized as follows:
1) A novel data-driven framework based on the reconstruction residual of autoencoders is proposed to evaluate voltage stability of power systems. The training of this method merely requires the data of secure state which is easy to collect, thus it is no longer subjected to the limitation that the practical data of insecure state is not sufficient.
2) To enhance the performance of the proposed method, a moving strategy for middle features is utilized to enhance the similarity between the features of testing data and the secure feature domain formed by the training data.
3) The proposed method is compared with other machine learning methods in imbalanced data manner and other types of autoencoders. The results demonstrate that the proposed method outperforms traditional algorithms for imbalanced data.
4) Multiple tests in different power systems are conducted. In addition to the classification accuracy, the computation cost and effects of measurement errors are analyzed empirically.
The remainder of this paper is organized as follows. Section II reviews the basics of voltage stability and introduces the background knowledge of autoencoders and the main principle. Section III introduces the entire methodology and improvements of the proposed method, including LSTM layers and feature moving strategy. Section IV introduces case studies, and Section V summarizes the research.
Long-term voltage stability, mostly suffering from load demand increments and unexpected changes of slowly acting equipment, involves the steady-state power system model that is described by an algebraic equation [
(1) |
where is the vector of state variables including nodal voltage magnitudes and angles observed from PMU at time ; is the loading factor portraying the gradual increase of load demand; and is the initial load level. In general, active load, reactive load and generator outputs at different buses increase at different rates, hence, the load growing model is commonly written as linear equations:
(2) |
where and are the active power and reactive power demand of load at a certain time , respectively; is the power output from generator at time ; and are the load demand of initial status; is the generator output in the basic case; and , , are multiplicative factors describing the growth rates of various variables as mentioned above.
The P-V curve, as shown in
(3) |

Fig. 1 Visualization of P-V curve.
where and are the VSM at current time and a base value, respectively. Otherwise, the operation state is judged as insecure state if , as shown by the red area in
The high-resolution and high-accuracy PMU measurements are used to assess voltage stability in this paper. PMU measurements contain nodal voltage magnitudes and angles, nodal injective active and reactive power, and branch currents. Voltage magnitudes and angles are selected for VSA because nodal voltages are the most representative measurement variables for voltage stability, and they are sensitive to the change of operation state. So the measurement vector at time is , where is the number of buses.
We use a split window that slides on measurement sequence of nodal voltage magnitudes and angles to collect a period of data , where is the length of the split window. Each two-dimensional data slice is used as a data unit to input into the proposed method.
Autoencoder is an auto-associative neural network that recovers the input data in the output from a compressed representation in low dimension. Autoencoder cascades an encoder and a decoder that are used to extract low-dimensional features and reconstruct the input data from these features, respectively. As shown in
(4) |

Fig. 2 A schematic diagram to visualize structure of an autoencoder.
where and are the outputs of the
(5) |
where and are the reconstruction data and the input data, respectively.
Autoencoders have been widely studied and modified to quite a few variants such as sparse autoencoder (SAE) [
For a data set which only contains the data of secure state, where is the total length of measuring time series in , each element is input into the encoder to generate a feature vector , where is the length of a feature vector. And then the decoder outputs the reconstructed data from . Thus, we have a set of features , and a set of reconstruction loss . Consider a data set of insecure state , where is the total length of measuring time series in . We obtain a set of feature vectors , and a set of reconstruction residuals , where is the feature vector obtained by . Autoencoders are essentially learning an identity function, but they first compress data into low-dimensional features and then reconstruct it. The low-dimensional features must lose certain information compared with the input data, hence, the reconstruction residuals cannot reach zero. By only inputting secure data for training, autoencoders learn how to recover secure data but are unfamiliar with the data of insecure state. Specifically, after proper training of autoencoders by mere secure data, the reconstruction loss of each element in the set is expected to become lower values. However, for the data of insecure state which is not included in the training set , autoencoders are expected to fail to recover it, i.e., the reconstruction loss is significantly greater. According to the reconstruction loss, a threshold is introduced to classify the secure and insecure states.
Remark: Traditional studies based on autoencoders unanimously attempt to enhance the representative ability of feature vectors [
Now we introduce the enhancement and details for the proposed methodology, including the explicit structure of the LSTM autoencoder and the moving strategy for compressed features.
Although we can implement VSA by using a simple autoencoder, the accuracy and robustness still need to be improved. The drawbacks are explicitly listed as follows.
1) VSA is typically a multi-variate time series issue that not only the spatial correlation between different measuring variables exists but also the temporal correlation is very critical [
2) The proposeed method is based on the principle that, for the data quite different from the training data, its reconstruction is expected to be very poor. Nevertheless, in practice, autoencoders may generalize well, i.e., insecure data can also be properly represented and recovered. Specifically, even though the data corresponding to secure and insecure state has completely different distributions, the insightful operation mechanism of the power system shown in (1) remains unchanged. Thus, it is likely that the autoencoder-based model also understands the operation pattern of the power system, and the reconstruction loss of the insecure data is still lower, which will cause the decrease of the effectiveness of the proposed method.
Aiming to solve the first issue of temporal correlation, a multiple-layer LSTM is embedded in this work before FC layers to construct the encoder. While in the decoder, multiple-layer LSTM is successively connected after FC layers, as the opposite of the encoder.
LSTM is an enhanced variant of RNN that not only connects output and input data but also connects current cell state with the previous state. LSTM has achieved state-of-the-art performance in many research areas associated with time sequence analysis. In the field of power system, it has been successfully applied for load forecasting [

Fig. 3 Structure of an LSTM cell.
The explicit calculation of an LSTM cell is listed as follows:
(6) |
where , , are outputs of the input gate, forget gate, output gate, respectively; and , , , , , , , , , , , are the weights and bias to be optimized. and are the sigmoid and tanh activation function, respectively. The input gate aims to extract favorable information from the input data and the hidden state of the last time . The forget gate is used to decide to drop or deliver the variables of the last cell state , and the output gate is used to construct the current hidden state .
To solve the second issue mentioned in Section II-A, we introduce a moving strategy for extracted features to make insecure features more similar to secure features, so that the reconstruction of insecure data is impeded. As shown in

Fig. 4 Visualization of two-dimensional extracted features by LSTM autoencoder with a total of feature points.
The features of insecure data contain two types: one is overlapped with secure domain , i.e., , the other is not overlapped with . If a feature of insecure data falls in the domain , it will yield a reconstruction data matrix that belongs to secure states. Thus, the reconstruction residual is relatively great and the insecurity is detected. At the opposite, it is very difficult to judge whether the features outside the secure domain (i.e., ) are capable of producing high reconstruction loss. The reason is that autoencoders are likely to possess promising generality to understand the insightful operation pattern of VSA, so that a small part of insecure data may also be recovered accurately. Faced with this dilemma, we propose a method to move these insecure features to secure feature domain, and thus, they are prone to yield very high reconstruction residuals.
The moving algorithm of extracted features should satisfy some rules as follows: ① the moving operation requires to be embedded into the LSTM autoencoder, and they are optimized jointly with the entire autoencoder model; ② not only are insecure features processed by our moving strategy but also secure features are moved in the identical way, for no prior knowledge about the testing data is available. Therefore, the moving strategy should change insecure features significantly with little impact on the features of secure data.
In this work, an enhanced -nearest neighbor (KNN) method is proposed to move extracted features to the center of a certain number of their nearest neighbors in the training data set. Given a pre-trained LSTM autoencoder and an incoming insecure data matrix , the extracted feature vector is obtained by the encoder. The distance between and a feature vector in is measured by the cosine similarity:
(7) |
where is the
(8) |
(9) |
where is the center of the selected features; and is the distance between and , which is used as the weights. The center is a linear combination of selected features, thus it must reside into the domain .
In order to establish a more flexible moving strategy, a temperature parameter is introduced to decide the degree of moving towards the center . The final point after moving is:
(10) |
where is the temperature parameter to control the moving degree. The principle of this moving strategy is that the influence for secure features by this moving is very little, because secure features in the domain must have numerous very adjacent points in the training data set, i.e., the nearest neighbors in the training data set of a secure feature are very close to this feature. Therefore, the distance between a secure feature vector and the center of its neighbors is not great. As for the feature of partial insecure data outside the feature domain , its nearest neighbors in the training data set are distant. Therefore, the moving strategy enables insecure features to move to the secure domain , which biases the data reconstruction by the decoder to improve the reconstruction loss.
This moving strategy is embedded in the middle of the LSTM autoencoder, as shown in the block diagram in

Fig. 5 Block diagram of proposed method.

Fig. 6 Entire framework of proposed method.
The operation of the moving strategy requires a set of secure features, and it is inevitable that the moving strategy will slightly change the secure features. Hence, we need to obtain a set of secure features before operating the feature moving strategy. A two-stage training process is designed to tackle this issue, as shown in

Fig. 7 Diagram of training and testing process.
However, there exist some limits for the application of the proposed method. At first, a certain number of PMUs are required, because autoencoders, or more general, learning algorithms need sufficient data to acknowledge the operation information of power systems and implement classification tasks. Secondly, for large-scale systems which yield a large amount of data, sufficient computation resources are required for offline training and online testing.
In this section, we prove the effectiveness and accuracy of the proposed method by numerous experiments. The first case elaborates the calculation process, including data generation, parameter settings and training. While the second case compares our method with other autoencoders based methods to verify the effectiveness of the proposed feature moving strategy. And the performances in different testing systems are demonstrated, including IEEE 30-bus, 57-bus, 118-bus systems, and European high-voltage transmission 1354-bus network [
This case aims to introduce the explicit process and configuration of our method by using IEEE 57-bus system.
The data pool contains three types of load changing directions. One is random increase rate of randomly selected growing load. We randomly select around 30% active and reactive load to increase at different rates. And the ascending rates are randomly assigned by a uniform distribution , where is the uniform distribution; and is a base value of corresponding load. The second increasing model is that a single load grows while other load keeps unchanged. The final way is that the load of a specific area ascends simultaneously, and every load increases in its respective rate which is randomly selected by w.r.t. the loading factor. The segmentation for IEEE 30-bus and 57-bus systems is based on the partition described in their documents, while the segmentation of IEEE 118-bus system is described in [
Now we list the explicit structure and hyper-parameters of our method as shown in
The indicator to evaluate the performance of our method follows the well-known index calculated by the classification precision and recall rate. The precision is defined as:
(11) |
where and are the number of correct and false classification results, respectively, if judging them into secure class. is to measure how many correct classification results if the proposed model recommends secure states. The recall rate Rc0 of secure state is:
(12) |
where is the number of incorrect results of the insecure class. Rc0 essentially measures how much secure data is correctly clustered w.r.t. the total number of secure data. Then the index is defined as:
(13) |
Similarly, the precision, recall rate and index for insecure data are defined as:
(14) |
where is the number of correct results if judging them into insecure class. The weighted index, which assigns different concerns to secure and insecure class according to their respective number of samplings, is used as the final indicator to evaluate the performance:
(15) |
Also, the weighted precision and weighted recall rate are defined as:
(16) |
Based on the training by pure secure data, the proposed LSTM autoencoder with moving strategy successively divides the secure and insecure data. As shown in

Fig. 8 Visualization of two-dimensional extracted features with moving strategy.

Fig. 9 Histogram of distribution of reconstruction residuals.
The criterion of reconstruction loss serving as the classification boundary is critical for the accuracy of the proposed method. We test different criterions and visualize the precision, recall rate and index of both secure and insecure data as shown in

Fig. 10 Changing trends of precision, recall rate and f1 index with criterion growing, using moving strategy or not. (a) Classification report. (b) Classification report without moving strategy. (c) Weighted average of classification report. (d) Weighted average of classification report without moving strategy.
Based on the simulation results mentioned above, several properties of the proposed method are revealed. At first, our method effectively implements VSA by pure secure data, which overcomes the disadvantage that insecure data or even unstable data is extremely rare. And through the proposed moving strategy, the classification accuracy is significantly improved. The best index without moving strategy is 0.9279, while the best index of our method embedding moving strategy is 0.9787, demonstrating the effectiveness and high accuracy of our method. The best criterion for classification also increases, since the feature moving strategy encourages higher reconstruction loss rate for insecure data, so secure data and insecure data are separated more clearly. The best criterion for simple LSTM autoencoder is 0.2574, while the best one for the proposed method is 0.2836.
Even if PMU has achieved enormous popularity for its high accuracy and resolution, measurement noise inevitably exists in the collected data of PMU. Higher measurement error will hinder the learning of the proposed method and may lead to overfitting. Therefore, it is infeasible to ignore the existence of measurement noise. To show the affect of measurement noise on our method, the classification results with different magnitudes of measurement noise are tested, as shown in
In this subsection, a large number of tests are implemented to compare the proposed method with other autoencoders using IEEE 30-bus, 57-bus, 118-bus systems, and European high-voltage transmission 1354-bus system. The criterions are selected by numerous tests as the first case. The detailed structure and hyper-parameters are listed in
The structure complexity of our model is changed according to different scales of testing systems. Larger systems such as European 1354-bus system and IEEE 118-bus system, which have installed 478 and 32 PMUs, respectively, have more complicated physical relationships that require more neural layers and more neural units to fit. For IEEE 118-bus system, we use three LSTM layers and five FC layers to construct the encoder and the decoder, and the criterion of reconstruction loss is 0.276. For European 1354-bus network, five LSTM layers and five FC layers are utilized, while the criterion of reconstruction residual is 0.5134. As for the IEEE 30-bus system containing 10 PMUs, only one LSTM layer and two FC layers are employed, and the criterion of reconstruction residual is 0.076. By comparison, the proposed method greatly outperforms other autoencdoer-based methods, including CNN-based autoencoder, SAE, and VAE, illustrating that the LSTM layer focusing on the temporal correlation is more beneficial for VSA. And the time of 500 operations is tested on Nvidia GeForce GTX 1080 (8G) GPU, as shown in
In this case, the proposed method is compared with other machine learning based methods, including one-sided support vector machine (OS-SVM), cost-sensitive decision tree (CSDT), and cost-sensitive random forest (CSRF). More information on these machine learning methods can be found in [
OS-SVM is an enhanced SVM algorithm to tackle the problem of imbalanced training data set. Compared with traditional soft SVM, it ensures absolute classification correctness of the main class by restricting its slack variables [
(17) |
where and are tuned parameters defining the boundary and supported vectors; and are a sampling vector and the corresponding label, respectively; and are the secure and insecure data domains, respectively; is a parameter governing the affecting level of slack variables; is the slack variable for ; and is the total number of sampling vectors in the data set.
The enhancement of OS-SVM is achieved by the third restriction for slack variables of secure data. If the classification is correct for , will be greater than zero; while an incorrect result leads to . Hence, the third restriction ensures for , i.e., OS-SVM is guaranteed to obtain correct results for secure data no matter how much insecure data is misclassified. OS-SVM is particularly suitable for extremely imbalanced data problem where the data of one class is very rare. For the VSA of binary classification, very limited insecure data but a large number of secure data is used for OS-SVM. This is different from the operation condition of the proposed method, which is totally free for insecure data. However, the comparison profoundly illustrates the accuracy and advantages of the proposed method.
Decision tree is a traditional and widely-investigated machine learning method which iteratively selects the most representative feature. In this paper, we employ classification and regression trees (CART), one algorithm of decision trees, to compare with the proposed method. To enhance the performance when processing imbalanced data set, a cost-sensitive loss function is used, which assigns different weights for classes according to their occurrence frequency.
(18) |
where and are the weights for secure and insecure data, respectively; is the total number of sampling vectors in the data set; and are the numbers of data points in these classes, respectively. , which denotes the number of classes, and these two weights satisfy . Since secure data is much more than insecure data in amount, is largely greater than . The loss function is defined as:
(19) |
where and are the estimation and the true class label corresponding to , respectively; is the loss function, which is the well-known cross-entropy in this paper. CSDT also requires a small set of insecure data, thus it is not totally free for insecure data.
Random forest is an ensemble learning method that aggregates numerous decision trees assigned by different input data, and then uses voting to obtain the final averaged result. It significantly mitigates the overfitting problem of decision tree and improves the performance. In this paper, we combine the weighted loss function with random forest, and use it to compare to the proposed method. More information on these machine learning methods can be found in [
For comparison, we use IEEE 57-bus system and the same preprocessing and data generation process. OS-SVM, CSDT, and CSRF require a small amount of insecure data, and more insecure data in training leads to more accurate classification. Thus, we use 15% insecure data, i.e., , , to implement OS-SVM, CSDT and CSRF, while the proposed method operates without any insecure data. The comparison results and data amount are listed in
This paper presents a reconstruction residual based VSA method that only requires the data of secure state. This work utilizes the well-known LSTM to form a spatial-temporal autoencoder, which is purely trained by secure data. Hence, the autoencoder is prone to produce lower reconstruction loss for secure data, while insecure data will encounter higher loss rate. To enhance the classification accuracy, a feature moving strategy for the middle features extracted from the autoencoder is proposed to enable insecure features to reside in the secure feature domain. This feature moving strategy guides the features of insecure data to move towards secure features, thus the insecure data is difficult to be properly recovered, i.e., higher reconstruction loss rate is obtained. The final reconstruction loss from the decoder is used as a proper indicator to detect the insecure operation state of VSA. Our method is particularly suitable for real validation, since no insecure data, which is very infrequent in practice, is required. In addition, it also has the priority of high accuracy and robustness for measurement noise. Further investigation could be considered to approximate directly VSM by only secure data. Some invariant features of operation states for VSA will be constructed, and insecure features will be adjusted to share the same changing trend with secure features. Thus, it is likely to estimate VSM by certain kind of principles of the extracted features from autoencoders.
References
Y. Qiu, H. Wu, Y. Zhou et al., “Global parametric polynomial approximation of static voltage stability region boundaries,” IEEE Transactions on Power Systems, vol. 32, no. 3, pp. 2362-2371, May 2017. [百度学术]
J. M. Lim and C. L. DeMarco, “SVD-based voltage stability assessment from phasor measurement unit data,” IEEE Transactions on Power Systems, vol. 31, no. 4, pp. 2557-2565, Jul. 2016. [百度学术]
Z. Wang and J. Wang, “A practical distributed finite-time control scheme for power system transient stability,” IEEE Transactions on Power Systems, vol. 35, no. 5, pp. 3320-3331, Sept. 2020. [百度学术]
X. Shi, R. Qiu, Z. Ling et al., “Spatio-temporal correlation analysis of online monitoring data for anomaly detection and location in distribution networks,” IEEE Transactions on Smart Grid, vol. 11, no. 2, pp. 995-1006, Mar. 2020. [百度学术]
H. Yang, R. C. Qiu, L. Chu et al., “Improving power system state estimation based on matrix-level cleaning,” IEEE Transactions on Power Systems, vol. 35, no. 5, pp. 3529-3540, Sept. 2020. [百度学术]
Y. Zhang, G. Pan, B. Chen et al., “Short-term wind speed prediction model based on GA-ANN improved by VMD,” Renewable Energy, vol. 156, pp. 1373-1388, Aug. 2020. [百度学术]
D. Q. Zhou, U. D. Annakkage, and A. D. Rajapakse, “Online monitoring of voltage stability margin using an artificial neural network,” IEEE Transactions on Power Systems, vol. 25, no. 3, pp. 1566-1574, May 2010. [百度学术]
B. Leonardi and V. Ajjarapu, “Development of multilinear regression models for online voltage stability margin estimation,” IEEE Transactions on Power Systems, vol. 26, no. 1, pp. 374-383, Jan. 2011. [百度学术]
S. Li, V. Ajjarapu, and M. Djukanovic, “Adaptive online monitoring of voltage stability margin via local regression,” IEEE Transactions on Power Systems, vol. 33, no. 1, pp. 701-713, Jan. 2018. [百度学术]
H. Su and T. Liu, “Enhanced-online-random-forest model for static voltage stability assessment using wide area measurements,” IEEE Transactions on Power Systems, vol. 33, no. 6, pp. 6696-6704, Nov. 2018. [百度学术]
Y. Fan, S. Liu, L. Qin et al., “A novel online estimation scheme for static voltage stability margin based on relationships exploration in a large data set,” IEEE Transactions on Power Systems, vol. 30, no. 3, pp. 1380-1393, May 2015. [百度学术]
D. N. Reshef, Y. A. Reshef, H. K. Finucane et al., “Detecting novel associations in large data sets,” Science, vol. 334, no. 6062, pp. 1518-1524, Dec. 2011. [百度学术]
Y. Zhang, Y. Xu, R. Zhang et al., “A missing-data tolerant method for data-driven short-term voltage stability assessment of power systems,” IEEE Transactions on Smart Grid, vol. 10, no. 5, pp. 5663-5674, Sept. 2019. [百度学术]
F. Yang, Z. Ling, M. Wei et al., “Real-time static voltage stability assessment in large-scale power systems based on spectrum estimation of phasor measurement unit data,” International Journal of Electrical Power & Energy Systems, vol. 124, p. 106196, Jan. 2021. [百度学术]
Y. Li and Z. Yang, “Application of EOS-ELM with binary Jaya-based feature selection to real-time transient stability assessment using PMU data,” IEEE Access, vol. 5, pp. 23092-23101, Oct. 2017. [百度学术]
V. Malbasa, C. Zheng, P. Chen et al., “Voltage stability prediction using active machine learning,” IEEE Transactions on Smart Grid, vol. 8, no. 6, pp. 3117-3124, Nov. 2017. [百度学术]
L. Zhu, C. Lu, Z. Y. Dong et al., “Imbalance learning machine-based power system short-term voltage stability assessment,” IEEE Transactions on Industrial Informatics, vol. 13, no. 5, pp. 2533-2543, Sept. 2017. [百度学术]
R. C. Qiu and P. Antonik, Smart Grid and Big Data: Theory and Practice. Hoboken: Wiley, 2015. [百度学术]
O. Kochan, “The technique to prepare a training set for a neural network to model the error of a thermocouple leg,” in Proceedings of 2019 9th International Conference on Advanced Computer Information Technologies (ACIT), Ceske Budejovice, Czech, Aug. 2019, pp. 101-104. [百度学术]
H. Yang, K. Ding, R. C. Qiu et al., “Remaining useful life prediction based on normalizing flow embedded sequence-to-sequence learning,” IEEE Transactions on Reliability. doi: 10.1109/TR.2020.3010970 [百度学术]
D. Gong, L. Liu, V. Le et al., “Memorizing normality to detect anomaly: Memory-augmented deep autoencoder for unsupervised anomaly detection,” in Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, Nov. 2019, pp. 1705-1714. [百度学术]
S. C. P. E. D. Dehaene and O. Frigo. (2020, Feb.). Iterative energy-based projection on a normal data manifold for anomaly localization. [Online]. Available: https://arxiv.org/abs/2002.03734 [百度学术]
C. Zhou and R. C. Paffenroth, “Anomaly detection with robust deep autoencoders,” in Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, Canada, Aug. 2017, pp. 665-674. [百度学术]
L. Wang, Q. Zhou, and S. Jin, “Physics-guided deep learning for power system state estimation,” Journal of Modern Power Systems and Clean Energy, vol. 8, no. 4, pp. 607-615, Jul. 2020. [百度学术]
S. Ryu, H. Choi, H. Lee et al., “Convolutional autoencoder based feature extraction and clustering for customer load analysis,” IEEE Transactions on Power Systems, vol. 35, no. 2, pp. 1048-1060, Mar. 2020. [百度学术]
J. Wang, D. Shi, Y. Li et al., “Distributed framework for detecting pmu data manipulation attacks with deep autoencoders,” IEEE Transactions on Smart Grid, vol. 10, no. 4, pp. 4401-4410, Jul. 2019. [百度学术]
H. Yang, R. C. Qiu, X. Shi et al., “Unsupervised feature learning for online voltage stability evaluation and monitoring based on variational autoencoder,” Electric Power Systems Research, vol. 182, pp. 1-11, May 2020. [百度学术]
P. Kundur, N. J. Balu, and M. G. Lauby, Power System Stability and Control. New York: McGraw-Hill, 1994. [百度学术]
T. Van Cutsem and C. Vournas, Voltage Stability of Electric Power Systems. Berlin: Springer Science & Business Media, 2007. [百度学术]
A. Rangamani, A. Mukherjee, A. Arora et al. (2017, Oct.). Critical points of an autoencoder can provably recover sparsely used overcomplete dictionaries. [Online]. Available: https://arxiv.org/abs/1708.03735 [百度学术]
D. P. Kingma and M. Welling. (2013, Dec.). Auto-encoding variational bayes. [Online]. Available: https://arxiv.org/abs/1312.6114 [百度学术]
A. Makhzani, J. Shlens, N. Jaitly et al. (2015, Nov.). Adversarial autoencoders. [Online]. Available: https://arxiv.org/abs/1511.05644 [百度学术]
L. Zhu, C. Lu, and Y. Sun, “Time series shapelet classification based online short-term voltage stability assessment,” IEEE Transactions on Power Systems, vol. 31, no. 2, pp. 1430-1439, Mar. 2016. [百度学术]
W. Kong, Z. Y. Dong, Y. Jia et al., “Short-term residential load forecasting based on LSTM recurrent neural network,” IEEE Transactions on Smart Grid, vol. 10, no. 1, pp. 841-851, Jan. 2019. [百度学术]
A. Gensler, J. Henze, B. Sick et al., “Deep learning for solar power forecasting: an approach using autoencoder and lstm neural networks,” in Proceedings of 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Budapest, Hungary, Oct. 2016, pp. 2858-2865. [百度学术]
M. S. Azman, Y. Isbeih, M. S. El Moursi et al., “A unified online deep learning prediction model for small signal and transient stability,” IEEE Transactions on Power Systems, vol. 35, no. 6, pp. 4585-4598, Jun. 2020. [百度学术]
M. Kaselimi, N. Doulamis, A. Voulodimos et al., “Context aware energy disaggregation using adaptive bidirectional LSTM models,” IEEE Transactions on Smart Grid, vol. 11, no. 4, pp. 3054-3067, Jul. 2020. [百度学术]
I. Goodfellow, Y. Bengio, and A. Courville, Deep learning. Cambridge: MIT Press, 2016. [百度学术]
D. P. Kingma and J. Ba. (2014, Dec.). Adam: a method for stochastic optimization. [Online]. Available: https://arxiv.org/abs/1412.6980 [百度学术]
X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, May 2010, pp. 249-256. [百度学术]
C. Josz, S. Fliscounakis, J. Maeght et al. (2016, Mar.). AC power flow data in MATPOWER and QCQP format: itesla, RTE snapshots, and PEGASE. [Online]. Available: https://arxiv.org/abs/1603.01533 [百度学术]
X. He, Q. Ai, R. C. Qiu et al., “A big data architecture design for smart grids based on random matrix theory,” IEEE Transactions on Smart Grid, vol. 8, no. 2, pp. 674-686, Mar. 2017. [百度学术]
J. F. T. Hastie and R. Tibshirani, The Elements of Statistical Learning: Data Mining, Inference, and Prediction. New York: Springer, 2009. [百度学术]
M. Wytock, S. Salapaka, and M. Salapaka, “Preventing cascading failures in microgrids with one-sided support vector machines,” in Proceedings of 53rd IEEE Conference on Decision and Control, Los Angeles, USA, Dec. 2014, pp. 3252-3258. [百度学术]