Transient stability is an important issue in power system planning, operation and control, as it is one of the most primary considerations in system security. In these applications, it is a regular routine to establish a description of transient stability boundary (TSB), no matter accurate or approximate, by performing time-domain simulation (TDS) for numerous operating points (OPs) and possible contingencies. Computational burden of TDS is especially heavy, if not unacceptable, after data-driven techniques have been employed to find out TSB. Thus, the key challenge is to improve the efficiency of obtaining sufficient TDS samples near TSB with representative scenarios for accurate boundary generation.
In general, most existing works are based on model-based approaches, like TDS , transient energy function  and extended equal area criteria , but with certain drawbacks of computational burden, model adaptability, or reliability, respectively. The advent of data-driven technique provides an alternative way to relieve such difficulties. A huge amount of transient simulation data are, however, required to construct a reliable stability boundary, as there is a lack of accumulated historical data with disturbances during daily operation, in terms of both variety and severity. Monte Carlo Sampling with TDS is applied in most existing works, changing all loads randomly within a predetermined range uniformly , , ,  or independently . However, generators outputs, in those works, are then scheduled by a certain rule (e.g., optimal power flow), to emulate practical system operations. Considering that the actual OP may not meet such rules, a good point set method  is proposed to improve the solution.
After realizing the low efficiency of the methods aforementioned, re-sampling based techniques are introduced in this area , ,  to enrich scenarios based on the existing generated data. Among them, binary search algorithm is employed in  to enhance the data set iteratively. Besides, importance sampling technique is also applied in an iterative framework ,  to identify the more important (or “informative”) region prior to TDS. Such efforts have been made to improve the quality of training data set for TSB.
Therefore, the goal of this work is to obtain sufficient TDS data near TSB with most critical scenarios and generate boundary by tracking current OP efficiently, which is necessary in dynamic security online monitoring and control. This is rarely addressed in the existing works, but exactly the key technology of data-driven based online security monitoring. Inspired by the adjoint sensitivity analysis (ASA) from dynamic optimization methods like , we propose to explore critical OPs close to the TSB, following the guidance of first-order derivative information of the specially designed transient index, so as to improve the information entropy of data samples. In addition, high-dimension variables clustering technique is also employed to find out the data relevance among the combinations of OPs and contingencies, in order to reduce the search space. As more data samples with representative OPs and contingencies accumulates, TSB can be constructed more accurately and efficiently.
According to this motivating idea, a data-driven transient stability boundary generation (DTSBG) framework for online security monitoring is proposed in this paper. As a continuation of the authors’ previous work of early terminating TDS for transient stability batch assessment , this paper is highlighted with the following contributions:
A critical data sampling framework with adjoint sensitivity based index, enforcing sampled data close to TSB;
A re-sampling method to fill more data in gap area of TSB generation, providing more samples across TSB;
A critical scenario selection strategy to identify the relevance of scenario set and relief computational burden.
Ii Transient Stability Boundary Generation
Ii-a Transient Stability Boundary Formulation
Stability boundary can be formulated as a set of constraints shown as follows:
where are the power flow equations, are the static operation constraints, are the initial condition equations, and are the transient constraints for the -th contingency. and are the upper and lower bound of the variable. While, as for variables in Eq. (1), are static variables of the given operation point, including active and reactive power of generators, and amplitude and phase angle of voltages. are the initial values of state variables. and are the time-variant state variables and operating variables respectively that describe dynamic response of the system after the -th contingency by solving the differential equations and algebraic equations shown in the following equations:
As a matter of fact, only the last equation in Eq.(1) is time-consuming while formulating the boundary because of the simulation part in Eq.(2). Therefore, in this paper, we only focus on the transient constraint or boundary of this type.
As for the transient stability inequality constraint, there are several ways to formulate. In this paper, we choose the absolute deviation of rotor angle with respect to the center of inertia (COI). It is shown in Eq.(3),
where , stands for the number of generators in the given power system, is the length of simulation time window while solving the DAEs defined in Eq. (2), denotes the absolute rotor angle of the -th generator under the -th contingency at time , and indicates COI which is defined as:
where denotes inertia time constant of the -th generator.
Ii-B Transient Stability Index and its Sensitivity
Since it is difficult to handle Eq. (3) directly, the constraint transformation technique is applied here to define a equivalent transient stability index as follows:
where is defined as:
In Eq.(6), indicates a constant, in order to distinguish the stable scenario from unstable one, and it is given by
Therefore, can be seen as an index to measure the dynamic performance of a given OP.
To guide the generated OPs near the stability boundary, it is of significance to obtain the first-order derivative information of the stability index . To do so, forward and adjoint sensitivity analysis are two major approaches. The latter one is preferred due to its lower computational burden. The first-order derivative information can be calculated as Eq. (8).
where is the Hamilton function corresponding to the transient constraint for the -th contingency, and can be defined as follows:
where and are the time solution to co-state equation:
Ii-C Critical Data Samples Sampling Strategy
In order to obtain more data samples close to the boundary, three main indices are required: one measures the distance between current sampled OP and boundary, the other two are the direction and step size guiding the current OP moving close to the boundary. In terms of distance measurement, we choose reflecting how stable/unstable the system is. In terms of direction, the first-order derivative information is employed. While, for step size, we design a variable step algorithm, in order to move the sampling OPs near the stability boundary as fast as possible. In that case, more critical OPs near the TSB can be sampled and simulated. Considering that larger step size can be set when the current OP is far from the boundary, while the opposite is smaller. Therefore, the step size is can be calculated as:
indicates a vector consisting with the maximum value of the controllable variables in the OPs, whileis the coefficient with respect to the stability index . Thus, the next sampling OPs can be calculated as follows:
After repeating this step until the new OP is found to be a critical one (close to the boundary), just as the route shown in Fig 1. However, some search processes cross critical area as the route considering highly nonlinear character of transient stability problem, a small trick like binary search is employed to handle such cases. Additionally, the new transient samples can also be positioned perpendicular to the gradient direction with small step size based on the existing samples near the boundary, as the route , to further enrich transient dataset close the boundary. Throughout the data collection process of these three routes, data samples with high information entropy are obtained for data-driven security assessment.
Ii-D Boundary Generation and Gap Area Re-sampling
After obtaining sufficient data samples using the strategy in previous subsection, TSB can be generated using data-driven algorithms, like decision trees, support vector machine, k-nearest neighbors algorithm, convolutional neural network, and etc. No matter what exact algorithm we choose, the essence of the boundary can be summarized as a complicated function with respect to the OPand contingency given as follows:
However, it is an undeniable fact that there might exist some sampling gap close to the security bound, although the sampling strategy improve the information entropy. Some parts of boundary may be inaccurate due to such problem. Thus, a re-sampling mechanism is proposed to find out new critical scenarios ignored in the previous search process. Essentially, this issue can be re-stated as an optimization problem that maximizes the distant to the existing sampling data closest to the stability boundary. Assume that data points have been generated, and new points are required in the sampling gap area. Thus, the optimization problem is:
where , , and .
As we can see from the equation shown above, it is a min-max optimization problem which is hard to solve. To simplify, an auxiliary variable is introduced into this optimization problem, realizing the chordal decomposition of minimum and maximum problem. Thus, the object function of Eq. (14) can be replaced by Eq.(15) shown as follows:
In fact, it is not necessary to find out the precise global optimal solution, considering that it is only a step for re-sampling. To reduce the required computational time, the duality gap can be preset a larger value. Meanwhile, more initialization points are selected before solving this optimization problem, considering the limitations of nonlinear optimization algorithm.
Ii-E Data Generation Termination Criterion
After focusing on data samples generation strategy, we should evaluate the termination criterion of the generation process. The auxiliary variable defined in Eq. (15) measures the distance vector between new sampling data and the existing data. Thus, the minimum element of can be utilized for data generation termination criterion and is defined as follows:
where is a preset termination threshold. A reasonable value is less than 1% of maximum controllable variables in OPs. Lower threshold improves the accuracy but sacrifices efficiency.
Iii Critical scenarios Selection
Iii-a Preliminary Search Space Selection
Normally, the proposed method aims at dealing with the changes of possible OPs and contingencies. However, the overall TSB is far too complicated to cover all possible scenarios, and thus we are trying to refresh TBS periodically by tracking current OP in time. In that case, the search space can be reduced significantly, and satisfied the limitation of online computational burden.
As shown in Fig. 2, the search space can be divided into two main dimensional categories: uncontrollable variables and controllable ones. In terms of the former ones, it mainly includes all kinds of loads, uncontrollable generators (like photovoltaic and wind generators). The exact value can be forecast day (hour or minute) ahead with the forecasting error less than 3% using state-of-the-art technique. As a result, variables in this part can be sampled randomly within a given small range. In terms of the latter category, it includes most normal generators. Considering the ramping constraint of each generator, the controllable variables are also be limited in a small range (although it is significantly larger than uncontrollable ones). Noted that the search area has a direct relationship with model periodic refreshing frequency. Overall, this process reduces the search space according to the latest scheduling OP.
Iii-B Critical Operating Point Selection
After determining the preliminary search space, the data relevance of all possible OPs are required to be analyzed. First of all, the contingency is assumed to be the same. Under such circumstances, we define a matrix with respect to the impact on different OPs under the same given contingency using stability index, that is, equals to:
where is the controllable operation variable (e.g., controllable generators), and are respectively the index and the dimension of variable . is the OPs, and are the index and the dimension of variable . denotes the index of the given contingency. is a matrix, and the -th element has column, and can be described as:
In order to measure the similarity of each OP, Spearman Correlation (SC) Algorithm is employed to classify all possible OPs into several clusters, based on the response of generator rotor angles for a given OP. To calculate the SC value, the matrix defined in Eq. (17) have to be converted into rank vector. For example, the -th row of indicates the derivative information of -th OP, and the controllable variable corresponding to the -th lowest value is assigned rank . Based on this definition, we can calculate the value of SC between the -th and -th OP by the following equation.
where and are the average value of rank vector with respect to the -th and -th OP, and are defined as follows.
After we got the SC value, spectral clustering technique is employed to classify the contingencies into several clusters. The most severe OP in each cluster can be regarded as the representative of others in the corresponding cluster.
Iii-C Critical Contingency Selection
Considering that the power system may undergo all kinds of contingencies, all possible disturbances have to be taken into account. As discussed in the previous subsection, grouping information has been obtained under the same contingency. In this subsection, therefore, we are trying to identify the representative contingencies under various OPs. If the grouping information is similar for two contingencies, only one needs to be analyzed in the next stage. Based on this idea, we employed adjusted rand index (ARI) to represent the relationship between contingencies and then cluster them into several sets. Given an OPs set with elements, and two contingency clusters of these elements, namely and . The overlap between these two contingency scenarios and are summarized in Eq.(21).
where, each entry (the -th row, -th column) denotes the number of OPs in common between and . Thus, the ARI between contingency and can be calculated as follows:
where denotes the combinatorial operator.
Similar contingency clustering process is carried out as critical OPs selection. So far, several critical contingencies are selected to reduce the computational burden.
Iii-D OP Matching and Critical Generator Selection
Although the aforementioned approaches have declined the number of OPs significantly, the real OPs vary from the ones in the day-ahead scheduling period, and the OPs matching mechanism is required. It is noted that the most critical generators (MCGs), in other words, the controllable generators that have the most significant impact on system transient stability, are almost the same in each OPs cluster. Therefore, MCGs can be determined once the new OP is matched into one certain cluster. We introduce the multivariate Gaussian model to do so. Suppose OPs under each contingency, the parameter and of a given cluster can be calculated:
Then, the possibility of a new OP in cluster is:
Lastly, the new OP is regarded to be in the cluster with largest possibility calculated in Eq.(24), and MCGs can be chosen accordingly.
Iv Online Security Monitoring Framework
Noted that transient stability boundary generation task is an NP hard problem, as the whole possible space of OPs are required to be discretized with small interval. Meanwhile, problem scale increases exponentially with the growth of power system interconnection, considering all possible discretized OPs together with all kinds of contingencies. It is definitely impossible to trace the whole TSB by classical brute force method, especially for power systems with more than 100 buses, before the advent of commercial quantum computers. Although gradient based sampling method is proposed in the previous section to improve the efficiency, search space still keeps the same. As a result, it is still hard to cover all critical scenarios, in order to generate an accurate boundary.
To relieve such difficulties, we introduce a new framework that only several controllable generators which affect stability most are included in the search space of a single scheduling period. Meanwhile, the search space can be further reduced by tracking the current or predicted OPs with relatively small refreshing time interval, in order to reflect the influence of other variables on TSB in time. Considering controllable generators which affect stability most vary from time to time according to different OPs and contingencies, scenarios clustering and matching technique is introduced in this step.
Based on such idea, we propose a new algorithm flow which is implemented as Fig. 4. It is observed that, in this figure, the whole process can be divided into three parts based on time scales: off-line precessing, periodic refreshing and on-line assessment. For the first stage, large amount of data with various scenarios are collected, and the most critical OPs and contingencies are selected, in order to make a preparation for most critical controllable generators selection process in the next stage. It is noted that this stage is executed off-line, and aims at reducing the search space. In terms of the second stage, the current OPs are matched with existing clustering result using multivariate Gaussian model. Meanwhile, critical controllable generators can be determined based on the matched scenarios, which significantly affect transient stability. After determining the critical controllable generators and scenarios, data are sampled and generated using the proposed gradient based method, to generate or refresh the accurate boundary. Meanwhile, data points are also re-sampled and generated until reaching the termination criterion. So far, an accurate TSB based on the current scheduling period has been obtained, and can be utilized in the online assessment stage. It is noted that the boundary is updated continuously by tracking the current OP, ensuring the accuracy of online assessment.
V Case Studies
Two typical systems with different scales are investigated in this section: IEEE 9-bus test system and NESTA 162-bus system. All cases are tested on a computer with Intel Core i7-4790 3.6GHz CPU, 16GB RAM unless otherwise specified.
V-a Visualize the Generation Process: IEEE 9-bus Test System
1) Test System and Configurations
In the first case study, the proposed algorithm is applied on a small system: IEEE 9-bus test system. The steady-state and transient parameters are referred in MATPOWER  and PSAT manual , respectively.
Considering that there are only 3 generators in this grid, it is still possible to visualize the generative process of TSB without dimension reduction and to verify all possible OPs that violate the security constraints of the given system. More specifically, it is highlighted in this subsection that the proposed algorithm is able to generate more data samples close to the TSB, in order to improve efficiency without compromising on accuracy of the boundary. Therefore, we measure not only efficiency but accuracy improvement visually compared with several existing methods.
In this study, all loads are assumed to change randomly and independently within 10 of its reference level. The contingency preset is initialed by a three-phase to ground fault at bus 5, and cleared after 0.2 seconds by tripping a line between bus 5 and 7. The stability performance is assessed using TDS and determined by maximum rotor angle difference. For comparison, accurate TSB is generated by brute force, that is, performing TDS at all possible OPs with discretized interval of 1MW.
2) Visualized Result of the Generation Process
As shown in Fig. 5(a), 20 initial OPs are randomly selected using Latin Hypercube Sampling method within the given output range of generators. Among them, 11 OPs with the ’’ mark are found in the infeasible area by static security check after solving power flow. The remaining 9 OPs are regarded as the initial seeds to generate the rest samples. Fig. 5(b) shows the process of getting close to the boundary using specially designed critical data sampling strategy introduced in Section II. It can be easily found that the step size varies according to the distance to the boundary. If the sample is getting closer to the boundary, the next step size becomes smaller. This ensures that the sampling near the boundary is sufficient, while traversing the rest less-informative part with fewer samples. Occasionally, some search processes cross the boundary due to the highly nonlinear nature of TSB, binary search is then employed to recover from such cases.
Additionally, the new transient samples can also be searched perpendicular to the gradient direction based on the existing samples near the boundary, as shown in Fig. 5(c). However, it can be observed that there exists the sampling gap on part of TSB. Therefore, more data samples in that area are required to generate a more accurate boundary. A rough boundary is generated as the red line and new sampled points are gotten in the gap area as green triangle markers in Fig. 5(d).
So far, new sampled points can be regarded as the new initial seeds and repeat the above procedures, so as to generate a more accurate boundary with the increase of data samples.
3) Gradient Information for Possible OPs
Fig. 6 shows gradient direction information for all possible OPs in static stable area, with discretization interval of 5MW. It is observed that almost all gradient direction arrows point to transient stability boundaries using the proposed transient index and algorithm. Therefore, it proves effective to generate more data samples close to the stability boundary using this specially designed index and method.
4) Numerical Results of the Proposed Method
Throughout the whole process of the data and boundary generation, the proposed method shows significant effect on efficiency without compromising accuracy. We compare the proposed method with four existing ones, summarized in Table I in details. For comparison more conveniently, we set a minimum accuracy requirement for all methods. Enabled by the proposed method, the data size required to reach 99.9% accuracy declines by nearly 80% and the time required reduces by 32%, compared with the most state-of-the-art method (importance sampling based method). It is highlighted that the proposed method shows better performance.
Meanwhile, the scatter plot of the data samples using different methods is illustrated in Fig. 7. We can easily distinguish the stability boundary through the last two plots (importance sampling based and the proposed methods) rather than the first three. Among them, the data samples generated by the proposed method are much closer to the boundary, and thus shows the superior performance.
|Latin Hypercube Sampling|
*All possible OPs (90,601 OCs) in this case are taken into consideration to evaluate the accuracy performance of each approach.
**The OPs in critical area can be seen as either stable or unstable points.
***In terms of existing methods, more samples result in higher accuracy and lower efficiency. For comparison, the minimum accuracy requirement is set to 99.9%.
V-B A Higher Dimension System: NESTA 162-bus System
1) Test System
In the second case study, the proposed method is applied to a larger and more complex power system named NESTA 162-bus system. Considering the large number of controllable generators and possible ”N-1” contingencies in this grid, computational burden increases geometrically in order to generate TSB. Therefore, we focus on the most critical scenarios selection under different circumstances, to ensure that the online computational burden is under control.
2) Most Critical OPs and Contingencies Selection
1,000 OPs are selected randomly according to the load prediction and dispatching plan within the given scheduling period. Meanwhile, 512 contingency, which is initiated by a three-phase-to-ground fault at any line close to bus of one end and cleared after 0.2 seconds by tripping the line, is also selected on all these 1,000 OPs. So far, 512,000 samples are selected based on day-ahead scheduling, to find out the most representative scenarios in the next scheduling period.
Various OPs under a single contingency are, firstly, to be clustered into several categories. The matrix as Eq.(17) is constructed, and employed for clustering. Fig. 8(a) illustrates the eigenvalue of normalized Laplacian matrix in spectral clustering of all 1,000 OPs under contingency #1. It is observed that the first 6 eigenvalues are relatively small, while the others are large. Therefore, cluster number is set to 6. Note that, the number of cluster varies from 2 to 8 for different contingencies. Fig. 8(b) shows the OPs clustering result under contingency #1. The colored matrix shows the correlation relationship between all different OPs under the same contingency. Darker matrix elements indicate weak correlation between the two OPs, and vice versa. Among them, most OPs belong to the first two clusters, and only small amount of OPs fall into other four clusters. Therefore, only 6 critical OPs are taken into account, since they represents all possible OPs within a given scheduling period. In other words, this reduces the number of OPs dramatically from 1,000 to 6.
After obtaining all grouping and critical OPs information for these 512 contingency scenarios, critical contingencies are also required to be identified. Similar algorithm is applied in this task with result shown in Fig. 8(c). As observed, it can be divided into 25 categories in total. So, only 25 contingencies are required to represent 512 preset contingencies.
Additionally, it is also necessary for us to analyze the efficiency of the clustering process, although it is carried out off-line. The time consuming of OPs clustering is only 2.8-3.0 seconds per each contingency. Considering this step is of natural parallel characteristic, asynchronous parallel algorithm can be employed here if necessary. While the time consuming for contingencies clustering is 24.8 seconds. In a word, it is time-effective to find out the most critical OPs and contingencies, compared to doing TDS for a huge amount of scenarios.
3) Test Results of Scenario Matching and Periodic Refreshing
In on-line operation stage, OP varies from time to time with different circumstances. Although huge amount of data samples are employed off-line, it is still impossible for almost all real OPs to match with the existing samples exactly. Considering that all clusters are difficult to be distinguished on a 2-D plane, we select three of them, shown in Fig. 9
, as an example to clarify this issue. In this figure, three selected OPs clusters are marked with different colors, together with the probability distribution contour plot. As seen in this figure, the dots with magenta asterisk mark belong to cluster #2, and MCGs in these scenarios are #1 and #12. In other words, these two generators are the key for operator to monitor and control in and near the current OP to prevent possible contingencies. Similar results are found for the other clusters except the critical generators. While encountering a new OP in real-world scheduling, multivariate Gaussian distribution probability result is utilized to evaluate the most possible cluster, and to determine MCGs. 1,000 new scenarios are evaluated and 96.1% of them obtain the same index of MCGs. Although the remaining 3.9% scenarios are not the most 2 critical generators, they still rank top 3 or 4. Considering the continuously updating in the scheduling period of the proposed method, it has little impact on the security assessment.
Besides, in order to evaluate the effect of periodic refreshing mechanism using the proposed algorithm in higher dimensional power system, we generate the dynamically updated TSB in a one-hour scheduling period with its load rate ranging from 0.8 p.u. to 1.1 p.u.(see Fig. 10). As shown in Fig. 11(a), all OPs in the search area (in the area surrounded by dotted line) are stable, when the load level is relatively low. At this stage, it only takes less than 10 seconds to generate the boundary within the search area, and make a conclusion that the state of the current OP is safe. It can be observed that, however, with the increase of load level, some OPs in the search area fall into transient unstable area shown in Fig.11(b). As a result, the operator are encouraged to reduce the outputs of generator #1 and #12 (see Fig.11(c)), while increasing others to some extent. By periodic refreshing the boundary, operators can adjust the current OP continuously, maintaining sufficient stability margin all the time. Even if the load level reaches its maximum as Fig.11(d), the OPs still in the stable area. More details can be referred in GIF animation111Results for more time sections in GIF animation format can be downloaded at http://genggc.org/files/YanTSB2020.gif.
It is noted that the time consuming to generate the TSB is less than 50 seconds (see Fig. 10 for more details) according to the length of boundary in search area. Thus, TSB can be refreshed every minute. Moreover, parallel technique can be employed in the future to generate data samples, because different search path (see route , and in Fig.1) is with the character of naturally parallel. Additionally, it helps in further reducing the refreshing interval to improve the hardware conditions in practical applications.
This paper has proposed a data-driven transient stability boundary generation framework for online security monitoring. In doing so, a critical data sampling framework and data gap area re-sampling mechanism have been proposed to accelerate the process of generating sufficient informative data samples near and across the boundary. Meanwhile, critical scenario selection strategy is developed to identify the relevance of scenario set and to further reduce the search space of high dimension power systems, enabling the possibility of periodic updating boundary tracking the current OP. The results of case studies illustrated that the proposed method reduces the computational burden of boundary generation process.
In sum, the proposed method offers advantages as follows:
Improving the efficiency of data generation with most critical scenarios;
Reducing the computational burden of boundary generation and periodic updating by tracking the current OP.
Enhancing the transient stability of the power systems by monitoring TSB and adjusting the current OP.
-  (1974-Jul.) Solution methods for transient and dynamic stability. Proc. IEEE 62 (7), pp. 951–958. Cited by: §I.
-  (1995-Nov.) Direct stability analysis of electric power systems using energy functions: theory, applications, and perspective. Proc. IEEE 83 (11), pp. 1497–1529. Cited by: §I.
-  (2010-Aug.) Decision tree-based preventive and corrective control applications for dynamic security enhancement in power systems. IEEE Trans. Power Syst. 25 (3), pp. 1611–1619. External Links: Cited by: §I.
-  (2019-Mar.) An online power system stability monitoring system using convolutional neural networks. IEEE Trans. Power Syst. 34 (2), pp. 864–872. External Links: Cited by: §I.
-  (2013-Nov.) Robust online dynamic security assessment using adaptive ensemble decision-tree learning. IEEE Trans. Power Syst. 28 (4), pp. 4089–4098. Cited by: §I.
-  (2011-Nov.) Efficient database generation for decision tree based power system security assessment. IEEE Trans. Power Syst. 26 (4), pp. 2319–2327. External Links: Cited by: §I.
-  (2017-Jul.) An efficient optimal control method for open-loop transient stability emergency control. IEEE Trans. Power Syst. 32 (4), pp. 2704–2713. External Links: Cited by: §I, §II-B.
-  (2014-Mar.) A systematic approach for dynamic security assessment and the corresponding preventive control scheme based on decision trees. IEEE Trans. Power Syst. 29 (2), pp. 717–730. External Links: Cited by: §I.
-  (2018-Mar.)(Website) External Links: Cited by: §V-A.
-  (2018-Mar.)(Website) External Links: Cited by: §V-A.
-  (2016-Sep.) Power system transient stability assessment based on big data and the core vector machine. IEEE Trans. Smart Grid 7 (5), pp. 2561–2570. Cited by: §I.
-  (1998-Aug.) Fast analysis of stability using eeac and simulation technologies. Powercon, pp. 12–16. External Links: Cited by: §I.
-  (2019-Jul.) Fast transient stability batch assessment using cascaded convolutional neural networks. IEEE Trans. Power Syst. 34 (4), pp. 2802–2813. External Links: Cited by: §I, §I.
-  (2018-01) Intelligent time-adaptive transient stability assessment system. IEEE Trans. Power Syst. 33 (1), pp. 1049–1058. Cited by: §I.
-  (2016-Jul.) Mining transient stability database for rule-based preventive control of power systems. pp. 1–5. External Links: Cited by: §I.