Detection of partial discharge patterns in hybrid high voltage power transmission lines based on parallel recognition method

View article
PeerJ Computer Science

Introduction

Hybrid high voltage power transmission lines (HHVPTL) composed of overhead lines and power cables are extremely substantial components in the power system and currently belong to the increasingly complex operating environment of the power system. HHVPTL gradually becomes the main artery of the power system (Venkatesh et al., 2020; Ding & Zhu, 2021). HHVPTL are prone to sudden failures due to long-term high voltage and high-current operation, resulting in external damage, equipment defects, and damaged insulation (Ragusa, Sasse & Duffy, 2021). Partial discharge (PD) monitoring is an efficient means to evaluate the insulation status of HHVTPLs and to assure their safe and stable operations (Houdai et al., 2022; Yamashita et al., 2018). The detection of the local discharge pattern of HHVTPLs is one of the important aspects of monitoring and diagnosis. However, due to the extremely complex compositional structure of HHVTPLs, the accuracy of PD diagnosis carried out manually is insufficient to meet the needs of power grid engineering. Therefore, pattern recognition methods with higher accuracy are needed.

The rest of the article is outlined as follows: the ‘Literature Review’ section is allocated to the literature review, Section ‘Feature Analysis and Extraction of the PD Signal for Hybrid High-Voltage Power Transmission Lines’ deals with extracting and assessing PDs’ features that can be used in the proposed algorithm, The proposed method and its fundamental notions are presented in Section ‘IDBO-BPNN-Based Pattern Recognition Model For Local Discharge Signals’. The experiments and comparisons are conducted in Section ‘Experimental Test’. The ‘Discussion’ section is allocated to the discussion. The research is concluded in the ‘Conclusion’.

Literature Review

To resolve the previously mentioned difficulties, more advanced emerging technologies are being applied to recognize the PD pattern of HHVTPLs. Currently, there are two main methods to recognize the local discharge pattern of power equipment. The first group is based on the manual identification of the collected PD data by experts, but the results are subjective. The other one is to train the local discharge pattern recognition model by implementing machine learning methods. Cheng et al. (2019) obtained the charge characteristics of cable joints by applying different levels of voltage and accomplished efficient identification of cable defects based on vector quantization theory and fast-matching algorithms. Zhang et al. (2022) proposed a novel approach to detect the PD pattern of power transformers. The ultrasonic signals caused by PD through the Fabry–Perot (F-P) ultrasonic probe sensing array are collected, and then the time-frequency representation is employed to convert the ultrasonic signals into a time-frequency matrix. Finally, the recognition of PD is performed through a modified ResNet-18 network. Huang et al. (2022) implemented an ultra-high frequency (UHF) sensor to acquire the phase-resolved partial discharge (PRPD) and phase resolved pulse sequence (PRPS) spectra and then designed a conjoined fusion network to accomplish the pattern recognition of PD. Chang, Chang & Boyanapalli (2022) designed a probabilistic safety assessment (PSA) method to identify the insulation status of power cable joints with PD by implementing a convolutional neural network (CNN) algorithm without measuring voltage signals. Yao et al. (2020) suggested that the data features to recognize the PD pattern of power equipment have high dimensionality and may contain redundant information. Thus, the random forest variance analysis (RF-VA) method was proposed to screen the optimal subset, which effectively reduced the dimensionality of the PD attribution set and enhanced the accuracy of PD defect detection.

However, the research on PD pattern recognition of HHVPTLs has been still relatively limited so far since the local discharge signal has uncertainty led by the special characteristics of HHVPTLs, which is primarily represented by the concentration of the main frequency and the poor effect of frequency analysis. Considering the actual environment of current power grid projects, PD pattern recognition is highly required for HHVPTLs. Therefore, the article proposes wavelet packet decomposition to analyze the time-frequency decomposition of PD signals. Moreover, the characteristics of the PD signal of HHVPTLs are deeply analyzed. So, a PD feature vector is built on this basis. Then, the improved backpropagation neural network (BPNN) utilizes them as inputs to train the algorithm. Eventually, a model that recognizes the PD pattern is obtained. Thus, the effective recognition of PD of HHVPTLs is realized, and the PD recognition technology of HHVPTLs is promoted. More up-to-date research presenting some practical modifications can be found in Faizol et al. (2023), Suren & Parvatha (2013), Hussain et al. (2023) and Shanmugam et al. (2023).

The highlights of the article are:

  1. Wavelet packet decomposition is employed to analyze the time-frequency decomposition of PD signals.

  2. The characteristics of the PD signal of the HHVPTL are represented by a feature vector that is used as the input of the improved BPNN algorithm that pinpoints PD patterns.

  3. An IDBO-BPNN algorithm is proposed to better classify the PDs.

Feature Analysis and Extraction of the PD Signal for Hybrid High-Voltage Power Transmission Lines

Feature analysis of PD signal

Considering the special attributions of the PD signal, the article proposes the utilization of wavelet packet decomposition (WPD) to perform time-frequency analysis on the PD signal of HHVPTLs and then analyze the features of the PD signal. When WPD is applied to analyze PD signals’ multi-resolution, the Hilbert space L 2 R = j z W j is employed to define. Then, L2(R) decomposition is performed based on distinct pairs of scale factors j to obtain the orthogonal sum of wavelet subspace Wj(j ∈ z) of wavelet function Ψ(t). The frequency resolution can be improved by binary frequency division of the wavelet subspace (Huang et al., 2023; Han et al., 2022). The new subspace Ω j n is employed to characterize the scale subspace Γj and the wavelet subspace Λj in a uniform way. Equation (1) presents this setting. Ω j 0 = Γ j Ω j 1 = Λ j .

Then, the decomposition of the new subspace Ω j n can be employed to unify the Γj+1 = Γj⊕Λj obtained by the orthogonal decomposition of L2(R)defined by Ω j + 1 0 = Ω j 0 Ω j 1 .

The subspace U j n is set to be the wavelet subspace of the function un(t) and the subspace Ω j 2 n is set to be the wavelet subspace of the function u2n(t). Then, the dual scale equation is constructed and defined (Zhang et al., 2019) by u 2 n t = 2 k z A k u n 2 t k u 2 n + 1 t = 2 k z B k u n 2 t k . where A(k) and B(k) represents the low-pass filter and high-pass filter coefficients, respectively, and are orthogonal to each other where Eq. (3) is transformed and rewritten by u 0 t = 2 k z A k u 0 2 t k u 1 t = 2 k z B k u 0 2 t k .

Equation (4) presents functions that are reduced to obtain the corresponding scale function φ(t) and the corresponding wavelet function Ψ(t), respectively in the multi-resolution analysis. The dual scale equation is constructed by Ψ t = 2 k z A k Ψ 2 t k φ t = 2 k z B k Ψ 2 t k .

By the equivalence of Eqs. (3) and (4), the condition of n ∈ Z is deduced. Then, Eq. (2) can be rewritten by Ω j + 1 n = Ω j 2 n Ω j 2 n + 1 .

The essence of the WPD is to further decompose the signal into two parts, namely, low frequency and high frequency, respectively. Then, s j n t U j n isset as follows: s j n t = r l j , n u n 2 j t l .

By Eq. (7), r l j + 1 , n is resolved to obtain r l j , 2 n and r l j , 2 n + 1 defined by r l j , 2 n = k α k 2 l r k j + 1 , n r l j , 2 n + 1 = k β k 2 l r k j + 1 , n .

Equation (8) is the WPD algorithm, which makes it possible to divide the PD signal into different WDs, which have the same bandwidth and are connected back and forth. The number of signal points contained is folded into half of the previous layer.

The PD data of HVPTLs from the Kaggle featured prediction competition is utilized. The dataset is normalized and categorized into two cases, namely, normal and faulty. The original signals of the two states and the results of the decomposition of the original signals implementing the WPD are presented in Fig. 1.

Typical PD signals of HHVPTL in different states.

Figure 1: Typical PD signals of HHVPTL in different states.

Figure 1 depicts that the PD signal composition of HHVPTLs is extremely complex, and it is difficult to guarantee higher accuracy by simply running the diagnosis manually. So, the extraction of the features of the PD signal is required to run a further diagnosis.

Feature extraction of PD signal

Figure 1 depicts that the PD signal energy of HHVTPLs is mainly concentrated in the fundamental frequency ranging from 0∼10,000 Hz. To swiftly grab the PD signal features of HHVPTLs, the article adopts the main frequency energy share of the PD signal obtained by the WPD as the core feature index. It adopts the assessment index of the conventional mechanical signal in Table 1 as the main attribute index of the PD signal to jointly establish the feature vector of the PD signal of HHVPTLs.

Table 1:
Main feature index.
Code Feature index Formula Code Feature index Formula
V1 Main frequency energy share V 1 = x 1 , n 2 / i = 1 7 n N x i , n 2 V 7 Minimum value x min = min x n
V2 Average value x = N n = 1 x n / N V8 Pulse Index I = x p x
V3 Standard deviation σ x = N n = 1 x n x 2 / N 1 V9 Skewness S = N n = 1 x n x 3 / N 1 σ x 3
V4 Root Mean Square x rms = N n = 1 x 2 n / N V10 Cliffness K = N n = 1 x n x 4 / N 1 σ x 4
V 5 Peak value xp = max|x(n)| V11 Spectral mean value F 12 = K k = 1 s k / K
V 6 Maximum value x max = max x n V12 Frequency center of gravity F 16 = K k = 1 f k s k / K k = 1 s k
DOI: 10.7717/peerjcs.2045/table-1

Since the frequency of the PD signal is primarily concentrated in the first decomposition node, the expression for the energy share of the main frequency V 1 is defined by V 1 = x 1 , n 2 i = 1 7 n N x i , n 2 where E0, N and xi,n denote the fundamental frequency energy, the length of the signal data sample, and the amplitude of each discrete point after the signal was reconstructed.

The final attribute vector of the PD signal of HHVPTLs that can be obtained consists of 12 total features denoted by V 1 through V 12.

IDBO-BPNN-based Pattern Recognition Model for Local Discharge Signals

BPNN

The essence of BPNN processes error terms between the actual output value and the predicted value by implementing the gradient descent algorithm. Namely, the derivative of the errors is resolved to tune the network weights and the threshold and the correction along the negative gradient of the derivative is continuously updated until the output error meets the preset requirements. When compared with the conventional algorithms, the advantage of the BPNN is that a deterministic mathematical model is no longer needed. The training stage is embedded in the model, and the corrected network can be obtained by tuning the parameters repeatedly (Zhan et al., 2022; Ou et al., 2022). The structure of BPNN is shown in Fig. 2.

The structure of BPNN.

Figure 2: The structure of BPNN.

The input layer in Fig. 2 contains n neurons that define the input vector, the hidden layer contains l neurons, the hidden layer contains m neurons, and the output vector is defined as the last layer of the network. The link weights between the input layer and the hidden layer are defined as wij, the link weights between the hidden layer and the output layer are defined as wjk, the threshold value of the neurons contained in the hidden layer is set to αj(j = 1, 2, ⋅⋅⋅, l), and the threshold value of the neurons contained in the output layer is set to βk(j = 1, 2, ⋅⋅⋅, l).

In the implementation of the BPNN, the training stage first determines network weights, and thresholds are continuously corrected during the training stage to transform the network into a prediction function. The steps are presented below.

Initialize the model parameters. The number of neurons in each layer is set according to the input and output vectors of the model, the link weights, the threshold of neurons in the hidden layer, and the threshold of neurons in the output layer are initialized, and the learning rate and the corresponding excitation function of the model are set.

Calculation of hidden layer output. The output vector H of the hidden layer is calculated based on the input vector of the model, the link weights between the input layer and the hidden layer, and the threshold of the hidden layer. H j = F u n c t i o n i = 1 n w i j x i α j , j = 1 , 2 , l .

Equation (10) represents the excitation function of the hidden layer.

Calculation of output layer output. Similar to , the network output vector O is calculated by employing the output of the hidden layer, the link weights of both the hidden layer and the output layer between each other, and the threshold value of the output layer. Equation (11) presents the output layer. O k = j = 1 l H j w j k β k , k = 1 , 2 , m .

Calculation of network error. The expected error Ek is obtained by calculating the difference between the expected output Ok of the model and the desired output Yk. Equation (12) presents the error term. E k = Y k O k , k = 1 , 2 , , m .

Correction of network weights. The link weights of the network are corrected according to the network error. w i j = w i j + ξ H j 1 H j x i k = 1 m w j k E k w j k = w j k + ξ H j E k where ξ denotes the learning rate, w i j and wij denote the updated weight at the next iteration, w j k and wjk represent the updated weight at the next iteration.

Correction of network threshold. The network threshold is corrected according to the network error, where α i and β k denote the threshold parameters to be updated at the next iteration. α i = α i + ξ H j 1 H j x i k = 1 m w j k E k β k = β k + E k .

Determine whether the output meets the termination condition and terminate the iteration if the preset score is reached, otherwise, return to .

Improvement of the BPNN based on orthogonal dyadic learning dung beetle optimization

Since the actual operation of the BPNN is mainly found by the network parameters, such as the number of nodes and connection weights, the final scores of the network parameters will have an impact on the BPNN’s results. To ensure that the BPNN has a good search performance and avoids falling into local optimums, an optimization algorithm needs to be introduced to optimize network parameters.

The dung beetle optimization (DBO) is the recently developed algorithm working on population intelligence optimization proposed in 2022. It is mainly derived from the habits of dung beetle life and has a robust search capability and high convergence efficiency. Thus, the DBO algorithm is employed to optimize BPNN in the manuscript.

DBO employs tentacle navigation when rolling to ensure that the dung ball keeps a straight line forward during the rolling process, and this behavior is needed in the simulation to permit the dung beetle to advance in the search space in the set direction and assume that light intensity affects the selection of dung beetle’s forward path. Then, the position of the dung beetle during the forward process can be defined by x i t + 1 = x i t + λ k x i t 1 + μ Δ x Δ x = x i t 1 x w o r s t

where t characterizes the number of iterations so far, xi(t) characterizes the position of the i-th dung beetle at the t-th iteration, and k ∈ (0, 0.2) characterizes the deflection factor, which is usually set to a constant value. λ taking either −1 or 1, and µdenotes a constant value in (0-1), xworst characterizes the local worst position, Δx representing mainly used regulation for light intensity.

When a dung beetle encounters an obstacle and cannot move forward, it needs to re-roll to reorient itself to develop a new route. To model the rolling orientation behavior, the new direction is resolved by utilizing the tangent function defined by. x i t + 1 = x i t + tan θ x i t x i t 1 .

In Eq. (19), θ ∈ [0, π] characterizes the angle of deflection and xi(t) − xi(t + 1) represents the difference between the anterior and posterior positions of the i-th dung beetle at distinct iteration cycles.

In addition, female, juvenile, and thieving dung beetles are defined in the population with position expressions as follows: x f d b _ i t + 1 = x b e s t + ϖ 1 x f d b _ i t l b + ϖ 2 x f d b _ i t u b x j d b _ i t + 1 = x b e s t + σ 1 x j d b _ i t l b c + σ 2 x j d b _ i t l b c x t d b _ i t + 1 = x b f + η ξ x t d b _ i t x b e s t + x t d b _ i t x b f .

To enhance the plurality of the dung beetle population, orthogonal contrastive learning is implemented to enhance it. Firstly, the orthogonal variation operator is introduced to perturb the optimal position in the population so that the joiner of the stealer masters from the variation solution of the optimal position of the finder, which can increase the population diversity with the ability to jump out of the local optimum, then Eq. (22) can be rewritten as x t d b _ i t + 1 = x b f + η ξ x t d b _ i t x b e s t + x t d b _ i t x b f .

In Eq. (17), = 1 , r < 0 . 5 1 , e l s e r¡ 0 is defined.

Contrastive learning is a commonly implemented strategy to jump out of the position found as the local optimal solution. In the detection of the DBO, the dung beetles in the optimal position move toward the worst solution, and the dung beetles in other positions move toward the optimal score. Although searching for the worst solution position can avoid falling into the local optimum to some extent, this is not conducive to the convergence of the population. In contrast, contrastive learning not only helps individuals to escape from the current position quickly, but also the contrastive position is more likely to have a better fitness score than the current worst position, so the contrastive learning strategy with these beneficial properties is introduced into the position update formula for rolling directional advance, and Eq. (19) is improved as follows: x i t + 1 = l b + u b δ x i t , f x i t f x i t 1 x i t + δ tan θ x i t x i t 1 , e l s e .

In Eq. (24): δ and β represent random numbers following N(0,1), lb and ub denote the lower and upper limits of the merit search, respectively.

Then, the overall optimization process of the IDBO-BPNN is presented as follows:

Step 1: Collect the local discharge data of HVPTLs, divide the training set and test set according to phase and state, and input the training set into the model for learning.

Step 2: Initialization of the BPNN network parameters, DBO population, and algorithm parameters.

Step 3: Solve all individual fitness scores based on DBO according to the objective function.

Step 4: Update the dung beetle positions and determine if they are out of bounds by utilizing the orthogonal opposition learning algorithm.

Step 5: Update the optimal position of the dung beetle and the fitness score.

Step 6: Repeat the above steps until the iteration limit is reached; if not, then skip to step 2; otherwise, output the global optimal solution and fitness score to the BPNN.

Step 7: Output the local release pattern recognition result of HHVPTLs.

Experimental Test

Data source and experimental environment

The data sources are extracted from the Kaggle featured prediction competition. The data are initialized to obtain 8,188 sets of normal PD signals and 8,188 sets of faulty PD signals, each with a measurement time of 20ms, containing 800,000 points, and a line operating frequency of 50 Hz. Each group of signals in the data set will experience a complete three-phase AC power supply operating cycle, and all three phases are measured simultaneously.

A method to recognize the local discharge pattern of the HVVTPL is proposed in the article. Tests were conducted on a computer whose experimental environment is shown in Table 2.

Table 2:
Experimental environment.
Type Name Content
Software environment Operating system Windows 10 64-bits
Test environment Matlab 2018B
Hardware environment CPU Intel(R) Core(TM) i7-9750H CPU @ 2.60 GHz 2.59 GHz
GPU NVIDIA GeForce GTX 1650
Memory 32 GB 2400MHz DDR4
DOI: 10.7717/peerjcs.2045/table-2

Experimental results

To test the accuracy and efficiency of the proposed IDBO-BPNN based on the data set composed of 12,600 subsets of data in the form of the PD signal of HVPTLs, normal and faulty data constitute half of the total number of the dataset. The training and test sets are divided according to the ratio of 7:2. The dataset is utilized as the input to the IDBO-BPNN to train the network. Then, the test set is employed to test the trained IDBO-BPNN network to check whether the constructed network recognizes the fault state of PD signals and calculates the accuracy. Then, the outcomes are compared with the computations of the backpropagation neural network (BPNN), support vector machine (SVM), Elman neural network ELM, deep belief network (DBN), particle swarm optimization-backpropagation neural network (PSO-BPNN) gray wolf optimization- backpropogation neural network (GWO-BPNN), genetic algorithm- backpropogation neural network (GA-BPNN), and dung beetle optimization-backpropagation neural network (DBO-BPNN) (Wu et al., 2021; Jiang et al., 2021; Dan et al., 2021). The obtained outcomes are shown in Fig. 3. While 0 indicates a normal PD signal, 1 designates a faulty PD signal.

Recognition results (A) BPNN (B) SVM (C) ELM (D) DBN (E) PSO-BPNN (F) GWO-BPNN (G) GA-BPNN (H) DBO-BPNN (I) IDBO-BPNN.

Figure 3: Recognition results (A) BPNN (B) SVM (C) ELM (D) DBN (E) PSO-BPNN (F) GWO-BPNN (G) GA-BPNN (H) DBO-BPNN (I) IDBO-BPNN.

Figure 3 depicts that the unoptimized BPNN has a relatively low accuracy of 71.8333% in the recognition of PD signals due to the limitations of its structure and the incapability of the iterative algorithm. In contrast, the recognition accuracy of machine learning algorithms such as SVM and ELM, the accuracy reaches only 76.5% and 74.8333%, respectively. In addition, the recognition accuracy of the DBN reaches 82.1667%. Due to their limitations, the above methods are not efficient in extracting features of PD signals, which leads to unsatisfactory recognition accuracy as well. After implementing the standard optimization algorithms such as particle swarm optimization (PSO), gray wolf optimization (GWO), and genetic algorithm (GA) to optimize the BPNN, the recognition accuracy of the obtained PD signal reaches 84.3333%, 86.5%, and 89.6667%, respectively, which is an important enhancement when compared with the unoptimized BPNN. The improvement of the BPNN based on the optimization algorithms is a better approach to attain higher accuracy.

After implementing the DBO to optimize the BPNN, the recognition accuracy of the DBO-BPNN achieves 90.1667%, which is higher when compared with all implemented algorithms, which indicates that the DBO algorithm optimizes the BPNN better when compared with other optimization methods. However, an accuracy error close to 10% is hardly acceptable in engineering implementations. Therefore, the recognition accuracy of the IDBO-BPNN obtained by implementing the DBO is improved with the orthogonal opposition learning algorithm to deal with this issue. Thus, the IDO-BPNN can achieve 95.5%, which is 5.3333% higher than the recognition accuracy of the DBO-BPNN and far more than other recognition algorithms. Therefore, PD signal patterns can be recognized more accurately and present practical performance.

Discussion

To further present the advantages of the IDBO-BPNN, the analysis of the computational efficiency of distinct algorithms is carried out. The mean computation time of distinct algorithms after 10 cycles is calculated and compared and the outcomes are summarized in Table 3.

Table 3 shows that the BPNN, as well as other machine learning algorithms, are relatively fast, basically not exceeding 50s, which is much lower than the 130.7s required for running the DBN since they are composed of simpler internal structures when compared to that of deep networks. When the optimization algorithm is implemented to improve the BPNN, the computation time increases significantly along with the improvement of recognition accuracy. Then the DBO can find the optimal solution of the BPNN faster than other optimization algorithms, so the time is significantly decreased. However, it is still higher than when the BPNN is not optimized. On the other hand, the balance between accuracy and efficiency has been achieved as much as possible. The IDBO, however, can help the BPNN reach the global optimum faster and improve the convergence efficiency of the network by implementing the orthogonal-contrastive learning algorithm, which is a better way to achieve a “win-win” situation in accuracy and efficiency.

Table 3:
Operational efficiency comparison results.
Algorithm Time (s)(Average of 10 operation cycles)
BPNN 36.4
SVM 48.6
ELM 30.2
DBN 130.7
PSO-BPNN 81.8
GWO-BPNN 97.9
GA-BPNN 101.6
DBO-BPNN 67.3
IDBO-BPNN 62.6
DOI: 10.7717/peerjcs.2045/table-3

In summary, the suggested IDBO-BPNN can efficiently and accurately detect the PD signal of HHVTPLs, which is suitable for the actual implementations and is a practical digital technology that can help improve the quality and efficiency of operation and inspection of power enterprises.

Conclusion

In the article, a parallel identification method of PD patterns in HHVPTLs is suggested based on wavelet analysis and IDBO-BPNN, resolving the energy share of the main frequency of PD signal by the WPD algorithm. The feature vector together with the time domain index and frequency domain index of the PD signal is constructed. The BPNN is improved by implementing the DBO algorithm based on orthogonal contrastive learning. Then, the suggested IDBO-BPNN can efficiently and accurately recognize PD patterns and diagnose PD faults. The experimental results show that the suggested IDBO-BPNN can balance accuracy and efficiency concurrently and has the potential to be broadly implemented in practical implementations, which can help the digital transformation of the smart grid.

The current research has some limitations as follows: 1. Only one data set is used to conclude even though the implemented dataset is a benchmark dataset. 2. The proposed method is compared with several robust machine learning methods, however, other deep learning models can be used to supplement more comprehensive outcomes. 3. The current level of accuracy is very high based on the DBO heuristic optimization algorithm. However, a comprehensive check could be done with similar algorithms to attain better results.

Supplemental Information

  Visitors   Views   Downloads