Open Access Paper
11 September 2023 Method of detecting power equipment data based on LSTM-LOF
Author Affiliations +
Proceedings Volume 12779, Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023); 127790H (2023) https://doi.org/10.1117/12.2688734
Event: Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023), 2023, Kunming, China
Abstract
In the production work of the State Grid, the power equipment is often in a high and low temperature environment. The timely detection of the fault of the power equipment is an important part of the stable development of the State Grid. This paper proposes a method to detect the quality of power equipment based on the measurement data of power equipment at room temperature. Based on the LSTM model, the working data of power equipment is predicted in the form of serialized time nodes during the experiment. Then, the feature data is extracted and dimensionality reduction according to the operating environment, and finally the anomaly detection is carried out according to the Local Outlier Factor (LOF) algorithm. The experimental results show that the prediction effect of the model prediction diagnosis results is ideal, and the LSTM-LOF combined with the model can be used to assist engineers in the quality diagnosis of power equipment.

1.

INTRODUCTION

As an important backbone enterprise related to national energy security and the lifeblood of national economy, State Grid Corporation undertakes the basic mission of providing safe and sustainable power supply for economic and social development. The power equipment plays an important role in the national grid. Generally speaking, to find the stability of power equipment, it is generally necessary to conduct artificial detection, measure its data, and then judge its stability according to its data. Three main issues need to be considered during data detection: estimation of current health status, prediction of future status and failure time, and determination of the impact of failure on system performance1. Anomaly detection specifically refers to the early detection of potential product problems 2

At present, most countries are encouraging the development of artificial intelligence and big data technology. In the direction of power equipment fault prediction, a large number of researches on the application of big data have been carried out and achieved fruitful results 3-7. Compared with the traditional empirical knowledge and basic mathematical model for power equipment quality prediction, some scholars have applied neural network to data prediction of power equipment8-9 Based on the method10, this paper proposes a power equipment data prediction detection method, and then carries out the envelope of machine learning algorithm. It can fully excavate the rule knowledge between data, which not only overcomes the large judgment deviation of traditional experience and the lack of rules in the existing knowledge base, but also avoids the data accuracy, high efficiency and excellent accuracy which are difficult to achieve with the traditional mathematical model due to the complex structure of power equipment. It has reference significance.

2.

LSTM DEEP LEARNING MODEL

In neural network prediction, for RNN, only the output of the last node can be accepted, and the weight of the whole layer network is shared, so gradient disappearance or gradient explosion is easy to occur. LSTM 11 model is a deep neural network model, which can carry out linear and nonlinear fitting of data features, carry out deep extraction of data features, and dig potential rules. It usually shows better prediction performance than traditional models in prediction tasks.

2.1

LSTM Network structure

The LSTM network layer is composed of cell units connected. The unit structure is shown in Fig.1,

Figure 1.

LSTM unit structure

00057_PSISDG12779_127790H_page_2_1.jpg

It can be seen that the repeated structure A of LSTM has four layers. The main line running through these units is line 2 above. The main structure of LSTM consists of three gates that control the increase and decrease of cell information, namely the forget gate, the input gate and the output gate.

Figure 2.

cell state

00057_PSISDG12779_127790H_page_2_2.jpg

Forget Gate

The forget gate is responsible for the reduction of information in cells, controlling what information should be forgotten.

Figure 3.

Forget Gate

00057_PSISDG12779_127790H_page_2_3.jpg

The forget gate formula is ft = σ(Wf · [ht–1, xt] + bf). After connecting ht–1 and xt it multiplicates a weight Wf and adds bias bf, The weight and deviation are the parameters that the network model needs to learn. If the size of hidden state (i.e., size of hidden Layer of neurons) is hsize then Wf is hsize * hsize. The value of hsize is set manually.

Input Gate

Figure 4.

Input Gate

00057_PSISDG12779_127790H_page_3_1.jpg

The input gate determines what information is saved in the cell and what information is added to the cell. The formula for the gate is as follows:

00057_PSISDG12779_127790H_page_3_2.jpg
00057_PSISDG12779_127790H_page_3_3.jpg

Input Gate (a Sigmoid function layer) and tanh layer, the two neural network layers will learn their parameters like the forgetting gate before.

Output gate

The output gate determines what information the cell needs to output. The output gate formulas are as follows:

Figure 5.

Output Gate

00057_PSISDG12779_127790H_page_3_4.jpg
00057_PSISDG12779_127790H_page_3_5.jpg
00057_PSISDG12779_127790H_page_3_6.jpg

Output gate also has its own weight parameters to learn.

The whole one-layer neural network is composed of several LSTM units, which send the original sequence (x1, x2, ⋅⋅⋅xn) into the network, through the LSTM network, and finally expand into the required dimensions from the full connection layer.

Figure 6.

LSTM network structure

00057_PSISDG12779_127790H_page_4_1.jpg

2.2

Evaluation

In this paper, MAE (mean absolute error) is selected as the evaluation model performance index, which represents the average value of the absolute error between the observed value and the real value. The calculation formula is:

00057_PSISDG12779_127790H_page_4_2.jpg

3.

DATA DIMENSION REDUCTION AND DETECTION MODEL

PCA algorithm is used to reduce the dimension of the predicted data, and the processed data is convenient for later detection, and then anomaly detection is performed according to local outlier factor (LOF) algorithm.

3.1

PCA

Principal component analysis (PCA) is a common linear dimension reduction method. The main way is to project the information in the high-dimensional space to the low-dimensional space through linear projection, and make the projected information retain the maximum original information and reduce the dimension of data.

Assuming that the original data set has m samples and n features, the original data set can be expressed as the following m×n order matrix, shown as Eq.6.

00057_PSISDG12779_127790H_page_4_3.jpg

Since the dimensional influence between features in the data set needs to be eliminated before calculating the matrix covariance, standardization needs to be carried out first, as shown in Eq.7.

00057_PSISDG12779_127790H_page_4_4.jpg

In Eq.8,

00057_PSISDG12779_127790H_page_4_5.jpg

Use the equation | λ-R|=0 to calculate the characteristic root of the matrix in Eq.7, and then use the characteristic root λ1 ≥ λ2 ≥ … ≥ λη ≥ 0 to calculate the orthogonal unitized eigenvectors e1, e2,…, en.

In the case of keeping the total variance of the data set unchanged, the contribution rate of the ith principal component is 00057_PSISDG12779_127790H_page_4_6.jpg (i = 1,2, …, n), and the cumulative contribution rate of the first q principal components after reduction is:

00057_PSISDG12779_127790H_page_4_7.jpg

After the original data matrix is processed by principal component analysis to reduce the dimensionality, the relationship between the principal components C1, PC2, …, PCs(s ≤ n) and the original data feature xi is:

00057_PSISDG12779_127790H_page_5_1.jpg

In Eq.10, aij and PCi are not related to each other, and 00057_PSISDG12779_127790H_page_5_2.jpg.

3.2

Local outlier factor

Local Outlier factor (LOF) is a density-based outlier detection method.

In sample set D, the distance between a sample point o and its nearest neighbor far from its kth is defined as the kth distance dk(o) of point o, and the Kth distance neighborhood Nk(o) of point o is defined as all sample points whose distance from point o is not more than dk(o), as Eq. 11 :

00057_PSISDG12779_127790H_page_5_3.jpg

In Eq. 11, d (o, o’) is the distance from point o’ to point o’, if d (o, o’) < dk (o), then the reachable distance dreach(o, o’) of sample point o is defined as dk (o),else for d (o, o’), As shown in Eq.12:

00057_PSISDG12779_127790H_page_5_4.jpg

Where, the nearest neighbor number k needs to be selected according to the actual sample size and sample distribution.However, sometimes sample distribution will affect the selection of outliers, so LOF uses locally accessible density to represent the density of sample point o, and defines the locally accessible density of point o as lrdk (o):

00057_PSISDG12779_127790H_page_5_5.jpg

|Nk(o) | is the number of sample points contained in the kth distance neighborhood of point o.

Define the average value of the ratio of the local reachable density of point o to the local reachable density of all points in the kth distance neighborhood Nk(o) as the local outlier LOFk (o), as shown in Eq.14:

00057_PSISDG12779_127790H_page_5_6.jpg

According to Eq.14, the lower the local reachable density of sample point o is compared with that of its kth nearest neighbor, the higher the value of LOFk (o) is, the greater the possibility that point o is an outlier.

4.

THE SIMULATION TEST AND RESULT ANALYSIS

Since the failure of power equipment occurs in high and low temperature environment, high and low temperature data is used as the detection standard. Figure 7 is the flow block diagram of the abnormal detection method of power equipment proposed in this paper. The whole process block diagram is divided into data prediction module and envelope detection module, in which the data prediction module mainly completes the data prediction. The envelope detection module completes the feature extraction of the prediction test data and the data dimension reduction processing. Finally, a data set that can be input into the local outlier factor model is formed to complete the learning and classification of input features.

Figure 7.

Flow chart of anomaly detection method

00057_PSISDG12779_127790H_page_6_1.jpg

The specific implementation steps are as follows:

  • Step 1: Preprocess the original room temperature data, input it into the LSTM model, and output the high and low temperature environmental data of power equipment products.

  • Step 2: Conduct data preprocessing on the data obtained in Step 1. Firstly, conduct data standardization operation, and then use principal component analysis to complete the data dimension reduction operation.

  • Step 3: The pre-processed data is taken as the data set, and the resampling method is adopted to sample the data set, and n new sub-data sets are obtained. The n sub-data sets are taken as the input to train n local outlier factor models, and n different base learners are obtained. The learning results of n base learners are averaged through the integration decision. The final classification results are obtained.

4.1

Model evaluation indicators

Anomaly detection is essentially a classification problem, the all samples are divided into two categories: abnormal state and normal state, the test using confusion matrix in table 1 for test results were described,

Table 1.

Confusion matrix

 Predicted to be abnormalPredicted to be normal
Actually abnormalTrue Positive(TP)False Negative(FN)
Actually normalFalse Positive(FP)True Negative(TN)

This experiment uses Precision (P) and Recall (R) as the algorithm evaluation indicators, and the calculation formula is as follows:

00057_PSISDG12779_127790H_page_6_2.jpg
00057_PSISDG12779_127790H_page_6_3.jpg

P represents the proportion of predicted anomalies and the number of actual or abnormal samples to the total number of predicted anomalies. The larger the value, the better the anomaly detection effect of the algorithm. R represents the number of predicted anomalies and actual or abnormal samples as a percentage of the total number of actual anomalies. The larger the value, the better the anomaly detection effect of the algorithm.

4.2

Data introduction

The experimental data are transformer data collected by Gansu Power Company of State Grid. Transformer products use load, local discharge, dielectric loss of insulating oil, polarization index, volume resistivity, sugar aldehyde content, ground current of iron core, insulation resistance of iron core, DC resistance of winding as detection characteristics. The high temperature of transformer products for environmental testing is 62 °C and the low temperature is -26 °C. The characteristic values of transformer products are recorded every 20 minutes. There are 50 kinds of sampling transformer products. In the test process of a transformer product, the monitoring data of its characteristic load is shown in Figure 8. The residual characteristics of the product are similar to the tendency of the load. It can be seen that the variation trend of product eigenvalue is highly correlated with the ambient temperature. Therefore, in this example, 9 characteristics of transformer products are analyzed as key parameters.

Figure 8.

Trend diagram of transformer load high and low temperature environment

00057_PSISDG12779_127790H_page_7_1.jpg

Transformer products will be tested at normal temperature, and the load curve of transformer products at normal temperature is shown in the figure 9 below:

Figure 9.

Ambient trend diagram of transformer load at normal temperature

00057_PSISDG12779_127790H_page_7_2.jpg

4.3

Analysis of experimental process and results

Due to different data units, the contribution to fitting is also different, so it is necessary to standardize the data to the same order of magnitude, and accelerate the convergence of the model. The standardized formula is shown in Eq.17.

00057_PSISDG12779_127790H_page_7_3.jpg

The experiment completed the programming calculation in Python3.8 environment, relying on the deep learning framework keras and TensorFlow.

Data processing

The mean and variance of normal temperature data were introduced into the network as cell state and hidden State, and the temperature sequence of the test was introduced into the network to predict the feature sequence in the test environment.

Each individual product data are extracted, united into the same data length, after contrast found at around 200, most data length hence take 200 for unification of data length, will integrate the different products, the result was a 3 d data, one of the first dimension represents the product number, the length of the second dimension represents the data, The third dimension represents the number of features. Use the first 40 products as a training set and the last 10 products as a validation set.

LSTM neural network selected the transfer function of Sigmoid function, the number of trained hidden nodes is 32, the optimizer uses Adam, the magnitude of network gradient update learning rate is 0.001, learning rate attenuation is 0.0002, the maximum number of network training is 20.

Effect display

Figure 10 shows the convergence of the loss function on the training set and verification set as the number of iterations increases. From the overall loss value, the loss value tends to converge to 0.09. After 15 training rounds, it began to converge.

Figure 10.

LSTM model loss convergence curve

00057_PSISDG12779_127790H_page_8_1.jpg

The final model prediction is compared with the real data (taking characteristic current reserve coefficient Product 1 as an example), as shown in Figure 11. The same can be used for data prediction of abnormal products, as shown in Figure 12.

Figure 11.

Data fitting comparison

00057_PSISDG12779_127790H_page_8_2.jpg

Figure 12.

Abnormal product data prediction

00057_PSISDG12779_127790H_page_9_1.jpg

It can be seen that the overall trend of the data is well fitted, which is also applicable to other parameters. It has certain practical significance and better guiding function for the power equipment experiment data prediction.

In order to reduce the redundancy of the features of the data set, speed up the training of the model, and reduce the amount of calculation of the anomaly detection algorithm, the principal component analysis of the features of the data set was carried out in this experiment. The contribution rate of each principal component is shown in the figure. It can be seen from Figure 13 that the cumulative contribution of the first two principal components is close to 50%. Therefore, in this paper, 9 transformer product features in the original data set are reduced to 2 principal component features, which can effectively reduce the actual operation of the subsequent LOF detection algorithm and improve the efficiency of transformer product abnormal detection.

Figure 13.

The principal component contribution rate

00057_PSISDG12779_127790H_page_9_2.jpg

The two-dimensional data after dimensionality reduction are input into the detection model of LOF to achieve successful visualization of the envelope surface. The classification effect of the high-temperature data model is shown in the figure. The normal products are inside the cochromatic envelope, and the abnormal products are outside the red envelope.

Figure 14.

High temperature data envelope detection

00057_PSISDG12779_127790H_page_10_1.jpg

Similarly, the predicted low temperature data can also be extracted, and then abnormity detection is carried out by LOF algorithm after dimensionality reduction. The envelope detection results are shown in Figure 15.

Figure 15.

Low temperature data envelope detection

00057_PSISDG12779_127790H_page_10_2.jpg

Combined with the high temperature and low temperature envelope detection maps, five abnormal points can be found. According to the product numbers, the abnormal product numbers are 45-50 respectively, and their detection data are quite different from other normal thruster products. According to the research of this experiment, the method can effectively detect abnormal products.

At the same time, in order to increase the accuracy, using the sampling method, 10 new data set, the 10 is the data set as the input of 10 local outlier factor model for training, get 10 different learning device, use the LOF-bagging algorithm and LOF algorithm for multiple tests, the final result shown in the table 2 below,

Table 2.

Comparative Results

 Average accuracy (P)Average recall (R)
LOF-bagging95.8%94.7%
LOF92.5%91.3%

The experiment shows that the algorithm proposed in this paper has higher prediction accuracy than the ordinary anomaly detection algorithm.

5.

CONCLUSION

In order to solve the problem of abnormal detection of traditional power equipment, this paper proposes a method of predicting and checking power equipment test data based on LSTM-LOF model. According to the previous normal temperature data, the LSTM model predicts the high and low temperature data in the actual working environment. PCA is used for data dimensionality reduction extraction, and finally LOF algorithm is used for data detection. This algorithm solves the problem of predicting power equipment components by relying on experience knowledge, and can assist engineers to detect anomalies. Meanwhile, this algorithm has higher prediction accuracy than ordinary anomaly detection algorithms.

Subsequently, with the increase of various data of power equipment, more base learners can be used for model training. Meanwhile, the integration strategy can be adjusted to weighted average, and the weight coefficients of different base learners can be dynamically adjusted in the process of model training, so as to obtain better anomaly detection results.

REFERENCES

[1] 

Wang, Y., Limmer, S., Olhofer, M., Emmerich, M. T. M., Back, T., “Vehicle fleet maintenance scheduling optimization by multi-objective evolutionary algorithms,” In 2019 IEEE congress on evolutionary computation, 442 –449 (2019). https://doi.org/10.1109/CEC.2019.8790142 Google Scholar

[2] 

Ordóñez, C., Lasheras, F. S., Roca-Pardiñas, J., de Cos Juez, F. J., “A hybrid ARIMASVM model for the study of the remaining useful life of aircraft engines,” Journal of Computational and Applied Mathematics, 346 184 –191 (2019). https://doi.org/10.1016/j.cam.2018.07.008 Google Scholar

[3] 

Wu, H. Y., Miao, W. W., Guo, B., et al., “Research on State Prediction Algorithm of Power Communication Equipment Based on Improved Decision Tree,” Computer and Digital Engineering, 49 (1), 17 –20,74 (2021). Google Scholar

[4] 

Gao, Y. X., Sun, S. Z., “Application of wavelet denoising combined with ARMA model in the prediction of failure rate of power equipment,” Journal of Inner Mongolia University of Technology, 38 (2), 122 –128 (2019). Google Scholar

[5] 

Liu, M. W., “Application analysis of improved association rule method in power equipment fault prediction,” Shandong Industrial Technology, 175 (3), (2018). Google Scholar

[6] 

Wu, H. Y., Chen, P., Guo, B., et al., “State prediction of power communication equipment based on attention mechanism and LSTM[J],” Computers and modernization, (), (10), 12 –16 (2020). Google Scholar

[7] 

Wang, C. B., Chen, G., Zhou, R., et al., “Fault prediction method of power equipment based on survival analysis model,” Power Big Data, 23 (5), 1 –8 (2020). Google Scholar

[8] 

Chen, H. J., Zhang, P., Jia, Y. F., et al., “Research on operating temperature prediction method of power equipment based on BP neural network,” Electronic world, (10), 40 –41 (2018). Google Scholar

[9] 

Yang, J. H., Liu, Y., Liu, J., et al., “Parallel F-LSTM Model and Its Application in Power Communication Equipment Fault Prediction,” Journal of Wuhan University, (3), 263 –268 (2019). Google Scholar

[10] 

Zhang, Y. F., Meng, F. Y., Wang, Y. Q., et al., “Research on temperature prediction method of power equipment based on improved LSTM,” Journal of Electronic Measurement and Instrumentation, 35 (12), 167 –173 (2021). Google Scholar

[11] 

Hochreiter, S., Schmidhuber, J., “Flat minima,” Neural Computation, 9 (1), 1 –42 (1997). https://doi.org/10.1162/neco.1997.9.1.1 Google Scholar
© (2023) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Zhicheng Ma, Jinxiong Zhao, Hong Zhao, Guangyuan Zheng, and Baohui Wang "Method of detecting power equipment data based on LSTM-LOF", Proc. SPIE 12779, Seventh International Conference on Mechatronics and Intelligent Robotics (ICMIR 2023), 127790H (11 September 2023); https://doi.org/10.1117/12.2688734
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Data modeling

Detection and tracking algorithms

Machine learning

Transformers

Education and training

Principal component analysis

Instrument modeling

Back to Top