logo

Anomaly detection of control rod drive mechanism using long short-term memory based autoencoder and extreme gradient boosting

NUCLEAR ENERGY SCIENCE AND ENGINEERING

Anomaly detection of control rod drive mechanism using long short-term memory based autoencoder and extreme gradient boosting

Jing Chen
Ze-Shi Liu
Hao Jiang
Xi-Ren Miao
Yong Xu
Nuclear Science and TechniquesVol.33, No.10Article number 127Published in print Oct 2022Available online 11 Oct 2022
37701

Anomaly detection for the control rod drive mechanism (CRDM) is key to enhancing the security of nuclear power plant equipment. In CRDM real-time condition-based maintenance, most existing methods cannot deal with long sequences and periodic abnormal events and have poor feature extraction from these data. In this paper, a learning-based anomaly detection method employing a long short-term memory-based autoencoder (LSTM-AE) network and an extreme gradient boosting (XGBoost) algorithm is proposed for the CRDM. The nonlinear and sequential features of the CRDM coil currents can be automatically and efficiently extracted by the LSTM neural units and AE network. The normal behavior LSTM-AE model was established to reconstruct the errors when feeding abnormal coil current signals. The XGBoost algorithm was leveraged to monitor the residuals and identify outliers for the coil currents. The results demonstrate that the proposed anomaly detection method can effectively detect different timing sequence anomalies and provide a more accurate forecasting performance for CRDM coil current signals.

Anomaly detectionCRDMLSTM-AEResidualsXGBoost
1

Introduction

The control rod drive mechanism (CRDM) [1] is the only movable equipment unit in the reactor pressure vessel and actuator of the reactor control and nuclear safety protection system. The reactor relies on the CRDM to pull the control rod to realize reactor start-up, power regulation, reactive compensation, and safe shutdown. A pressurized water reactor (PWR) nuclear power plant (NPP) adopts a magnetic lifting drive mechanism that relies on coil power to realize control rod action. The control rod moves step-by-step through the alternating action of the three sets of coils in the driving mechanism [2]. Each step requires close cooperation between the three sets of coils and hook claw. If a slight error occurs, failure to pull, drop, or lift may occur. Therefore, it is of significant value to investigate how to improve the CRDM abnormal detection performance.

Recently, the state parameters used for abnormal detection of control rods are predominantly vibration signals, structural noise and inductance values. In actual control rod operation, each action will produce characteristic acoustic signals related to the moving parts. The operation [3] and wear [4] of the CRDM can be judged using acoustic-sensitive sensors and instruments, which provides an effective means for CRDM performance monitoring and fault analysis. Caylor et al. [5] proposed an abnormal detection method of the CRDM based on the principle of structural noise detection. The structural noise was obtained by the sensor, and the signal was analyzed and processed to determine the fault. Fault diagnosis based on structural noise provides a new fault-detection method for the monitoring and maintenance of CRDM in nuclear reactors. In addition, because the air-gap magnetomotive force is proportional to the equivalent length of the flux path, the air-gap length changes significantly when the actuator is eccentric [6], resulting in a change in the coil flux and the inductance value of the control rod coil is affected [7]. Xie et al. [4] proposed a new method for the state detection of CRDM action components by on-line measuring the coil inductance value, and realized state detection of CRDM action components. Ling et al., [8] proposed a model feature-based flow field fusion simulation method applied in CRDM. The research obtained a fusion simulation result of the CRDM through the hierarchical structure of the function-motion-action method, which presented a new concept for CRDM anomaly detection. The aforementioned research on CRDM anomaly detection can only judge the action component fault, and the CRDM mainly transmits the signal through the current during the operation process. To accurately obtain the characteristic information of the operation process, it is necessary to study the operation current signal detection method.

An increasing number of applications involving machine learning in the nuclear energy field are being implemented owing to its advantages of rapidly processing large amounts of data, analyzing and extracting effective information, etc. Chen et al. [9] designed a neural network accelerator to efficiently extract amplitude and time information from row-by-line pulses. Ma et al. [10] used the LSTM model to achieve optimal estimation of nuclear pulse parameters. However, development in machine learning has been slow in nuclear energy anomaly detection. In other fields, it has been widely used for conditions monitoring [11], fault diagnosis [12], and anomaly detection [13]. The effective perception of periodic and nonlinear change characteristics of time-series data has become a key factor in the effectiveness of various fault detection and diagnosis models. Traditional anomaly detection algorithms, such as Gaussian anomaly detection [14, 15], SVM [16, 17], KNN [18, 19], PCA[20, 21] and isolation forest [22, 23] do not consider anomaly detection in time series. Reference [24] used a competitive neural network to establish a prediction model of time series using historical data and judged whether there was an anomaly by comparing the difference between predicted and actual data. Yan et al. used this concept to establish a time-series state model, and clustering to map the time series to different discrete states, and used the statistical state transition probability to mine outliers. However, the discrete state modeling method loses a lot of data information and is overly reliant on the clustering results. Ergen et al. [25] used a new recurrent neural network model LSTM to model the sequence. This model can capture the relationship between sequences well, however, it is too complex to meet real-time requirements. Moreover, owing to the gradient disappearance problem, the modeling effect of long sequences is poor. Currently, most methods compare the difference between the predicted and actual data to determine abnormality. The hypothesis test data were subject to a normal distribution [26], and the 3 σ criterion was used for discrimination. This type of method ignores the real distribution of data and errors in small datasets.

This study proposes an anomaly detection algorithm based on LSTM-AE and XGBoost [27] for the coil current of a CRDM. Hence, this research realizes CRDM anomaly detection under various abnormal conditions and accurately obtains CRDM characteristic information by monitoring the CRDM coil current signal. The proposed method utilizes the LSTM-AE neural network to automatically extract the nonlinear and sequential features of the coil currents and builds a normal behavior model to reconstruct the residual under abnormal conditions. Long series and periodic historical current data can be handled and predicted using the LSTM-AE model. The XGBoost classification algorithm is employed to identify outliers in the reconstructed errors and avoid defects in hypothesis testing. The long series and periodic CRDM coil currents can be accurately predicted and detected by the generalization and classification ability of the proposed method.

2

Working principles of CRDM

The CRDM is primarily composed of a drive shaft, coil, armature, gripper, and other components [28]. As shown in Fig.1. Each control rod contains one lift coil (LC), movable coil (MC), and stationary coil (SC). The three groups of electromagnetic coils turn on and off alternately according to a certain time sequence, and the armature action is lifted, moved, and stabilized to achieve the rise or fall of the control rod. Therefore, the coil current size, on-off time sequence, and current sequence response of the hook claw action are extremely important for the mechanism action.

Fig. 1
(Color online) Structure of CRDM.
pic

According to the lifting and insertion control signals shown in Fig.2(a) and 2(b), the control rod is lifted or inserted into the reactor core through the CRDM step motion sequence. The CRDM coil operates at three voltage levels: high, low, and zero. After the coil was electrified, a low voltage was generated on the SC to ensure the control rod remained in a fixed position. A high voltage was applied to the LC. to start the motion of internal magnetic components. This initial motion causes the internal clearance between the magnetic components to close, thereby lifting the control rod. Subsequently, a low voltage was applied to the coil to ensure the gap remained closed, allowing the locking device to connect. The current was removed from the coil to allow gap opening and load transfer to the locking device. This control sequence enables the CRDM to move the control rods during and insertion actions. Figs.2(c) and 2(d) show the coil current curve during the lifting and insertion actions, respectively.

Fig. 2
Timing sequence and current waveform for lift and insertion motion.
pic

As shown in Fig.3, the current curve of a SC is divided into six stages described as follows.

Fig. 3
Stationary coil current.
pic

1) Zero current stage: at this stage, the coil current is zero, and the coil is in an open state.

2) Initial electrification stage: at this stage, the coil current begins to increase. Because the coil is an inductive load, the current slowly increases. After which, the electromagnetic force gradually increases. When the coil current does not increase to a certain degree, the magnetic force is small and does not produce action.

3) Back ditch stage: when the electromagnetic force increases to a certain extent with the current, the gripper moves and the current continues to increase. The magnetic gap on the upper magnet decreases, and the lifting force increases. However, the position of the coil changes during the action, therefore the corresponding magnetic flux becomes smaller, and the coil inductance becomes larger, which leads to a decrease in current and produces a return channel during the dynamic current change.

4) When the action is complete, the position of the hook claw and inductance of the coil no longer change, and the coil will form a transient steady state. Under this state, the current continues to rise to the inductive load to charge the current slowly until it stabilizes and reaches its peak.

5) High-level stage: at this stage, the coil is in a clamping state.

6) Low-level stage: the control signal is that the low-level current slowly drops into the low current, and the coil is in the clamping state.

Abnormal data contain types of anomalies common in CRDM operations. Figure 4(a) shows the abnormal current size, the fault of the CRDM current control circuit causes the output current to deviate from the set value. Excessive current may cause the coil to overheat, and a current that is too small may cause the armature to be unable to move. Figure 4(b) shows the non-action point anomaly. If there is no pit in the rising section of the current, it can be confirmed that the corresponding claw does not act. This anomaly can easily lead to stick-slip. The anomaly shown in Fig.4(c) represents the backward movement of the action point. The moving step requires seven actions and cohesion between the stationary gripper, movable gripper, and lift armature to cooperate closely. A slight error may lead to pull, drop, or lift failure. Fig.4(d) shows slow current rise. As a result, the action point of the armature lags behind, and it is easy to pull or slide the rod and scratch the drive shaft.

Fig. 4
CRDM coil current anomaly type. (a) abnormal current size. (b) backward movement of the action point. (c) non-action point anomaly. (d) slow current rise.
pic

Based on the aforementioned analysis, the CRDM coil current information contains the control rod action information. Accurate real-time extraction of coil current characteristics is the key to achieve CRDM abnormal detection. A schematic of the proposed anomaly-detection method is shown in Fig.5. The normal coil current operation data were used as input for the LSTM-AE model for feature learning, and the normal behavior model(NBM) was trained. The NBM represents the dynamic relationship between variables. When the data were normal, the output of the model was the same as that of the actual result. When an abnormality occurred, the output of the model differed significantly from the input data. When the NBM model was established, the abnormal data were entered into the NBM test to obtain the reconstruction error, and the XGBoost algorithm was used to detect the outliers. The detailed process of anomaly detection using the proposed LSTM-AE and XGBoost approach is described below.

Fig. 5
(Color online) Schematic diagram of CRDM anomaly detection using LSTM-AE and XGBoost.
pic

First, the coil current signal of the control rod drive mechanism during actual operation was collected, and the data were cleaned to obtain normal data. Second, based on the LSTM-AE algorithm, the NBM model was established using normal data. Finally, the abnormal coil current signal was used as the input of the trained normal behavior model, and the residual of the model reconstruction was obtained. Abnormal point detection of the control rod coil current was completed as the input for the XGBoost algorithm.

3

LSTM-AE neural network

3.1
The LSTM network

The LSTM model is a type of time-cycling neural network that contains three parts: the input layer, output layer, and hidden layer. LSTM can effectively address time-series learning problems. Its structure is illustrated in Fig.6.

Fig. 6
Architecture of a long short-term memory (LSTM) unit.
pic

LSTM has three gated structures to control the state of the cell, namely, the forgetting gate, the input layer gate, and the output layer gate. The input and output gates control the memory cell input and output activation. The forget gate updates the state of the cell. The storage cell update is given by the following equation: ft=sgmoid(wxfxt+whfht1+bf) (1) it=sgmoid(wxixt+whiht1+bi) (2) gt=tanh(wxgxt+whght1+bg) (3) ot=sigmoid(wxoxt+whoht1+bo) (4) st=ftst1+itgt (5) ht=ottanh(st) (6) where i indicates the input gate, o is the output gate, f is the forgetting gate, ft indicates the state of the cell at the previous time, st indicates the memory cell, ht indicates the final output, and represents the element-wise multiplication. wxf, whf and bf are the neuron weights of ft. wxi, whi and bi are the neuron weights of it. wxg, whg and bg are the neuron weights of gt. wxo, who, bo are the neuron weights of ot.

3.2
The LSTM-AE neural network

The LSTM-Autoencoder is divided into an encoder and decoder. The input data features are learned by the encoder, and then the reconstruction errors are calculated by the decoder Set the input data sequencex1,x1,x2,...,xn can be defined as {Xi}i=1n Where n indicates the size of the data window. For each input data Xi= {xi1,xi2,...,xin, the state vector of the encoder hidden layer corresponding to column i is as follows: Xgti=k(Wxti+Rh(t1)i) (7) h(t-1)i is the output state vector of the first decoding unit at t-1. xti denotes the input vector. W, R is the m×d-m×m-order coefficient weight matrix. Function k(.) is typically set as the ’tanh’ activation function. By entering each column vector in xi in the encoder section, the output is as follows: hti=kφenc(xti,h(t1)i) (8) For the output of the i encoding unit at time t, φ is the parameter set for the encoding part. kφenc(.) is typically set as the ‘tanh’ activation function. The output sequence {hti}i1ni was obtained using the coding part set. The output sequence was pooled using Eqs. (9), and (11). hi=j=1nihtini (9) hi=ht,ni (10) hi=MAX{hti}i=1ni (11) where j denotes the number of rows of hti. After the pooling operation, hi enters the encoder part, and the input can be reconstructed as follows: h^ti=kφdec(hi,h^(t1)i) (12) x^ti=ρ(h^ti) (13) where x^ti denotes the reconstructed data. h^ti is the implied state vector for the decoder and kφdecc(.) and ρ(.) are typically set to the ‘tanh’ function. Finally, the LSTM-AE model can be determined by minimizing function aaa. The LSTM-AE network structure is shown in Fig.7.

Fig. 7
Structure of LSTM-AE neural network.
pic
3.3
XGBoost

The XGBoost algorithm integrates several CART trees. XGBoost is an optimized distributed gradient-lifting algorithm. The main advantages of this algorithm are its efficiency, flexibility, and portability to other systems. The superior performance of XGBoost in supervised machine learning is the reason for choosing it to train the coil current anomaly classification in this context.

The XGBoost structure is shown in Fig.8. In each algorithm iteration, residual errors are used to modify the previous classifier and optimize the loss function, which is given by J(α)=L(α)+Ω(α) (14) α represents the trained parameter, L represents the loss function, Ω represents the regular term, and the output of the model y^i is voted on by the set F of the k tree: y^i=i=1kfk(xi),fkF (15) The objective function is converted at t time to J(t)=i=1nL(yi,y^i)+k=1tΩ(fk) (16) where n denotes the predicted amount. y^i represents: y^(t)=k=1tfk(xi)=y^(t1)+ft(xi) (17) The regularization term of the decision tree Ω (f k) is expressed as follows: Ω(fk)=γT+12λj=1Twj2 (18) where γ denotes the complexity of the leaf. T indicates the number of leaf nodes. λ denotes the penalty factor. w denotes the fraction vector. Assuming that the loss function is the mean square error, the objective function can be obtained as follows: J(t))i=1n[giwq(xi)+12(hiwq(xi)2)]+γT+12λj=1Twj2 (19) q(.) represents a function that assigns data points to the corresponding leaf nodes, gi represents the first derivative of the loss function, hi represents the second derivative. The loss function can be expressed as the sum of the loss values of leaf nodes. J(t)j=1T[(iIjgi)wj+12(iIjhi+λ)wj2]+γTGj=iIjgi,Hj=iIjhi (20) Ij represents all the leaf node data. Therefore, the optimization of the objective function is equivalent to determining the minimum quadratic function.

Fig. 8
Structure of XGBoost.
pic
4

Results

The experimental data were obtained from nuclear power plant 1-4 megawatt-class pressurized water reactor nuclear power (NPP) units. In the cold CRDM requalification test of this NPP, we collected and stored the LC, MC, SC currents, and vibration data through the RGL system. In this study, we conducted an anomaly detection investigation on the CRDM with an SC current as the research object. Based on the staff operational reports during the test, this experiment collected 20 cycles of normal coil current data as a sample set, with each cycle lasting ing 792 ms. During the construction of the LSTM-AE model, 80% of the sample set (16 cycles of normal coil data) was used as the training set and 20% (four cycles of normal coil data) was used as the test set. Four types of typical anomalies appeared in the control bar card shibboleth report during the test process. In this study, four types of anomalous coil currents were selected as the validation set of the LSTM-AE model, where each type of anomalous data contained four cycles. To enhance training efficiency of the deep learning model, all data were normalized between 0 and 1.

4.1
NBM based on LSTM-AE

In this case, the NBM of an induced CRDM was established using an LSTM-AE neural network with an SC current as the model input. Normal operational data were obtained based on the operational log. Twenty cycles of observations were obtained from the sample. All datasets were standardized, and the ratio of the training and test sets was 8:2. To demonstrate the effects of the established NBM model. Figure 10 describes the reconstruction result of NBM based on the LSTM-AE model. It can be observed that the curves of the reconstructed sample and that of the raw sample are almost coincident. This indicates that the established NBM can reconstruct the normal samples well.

Fig. 10
Reconstruction precision of NBM based on LSTM-AE model.
pic

In the LSTM-AE training process, the time step and number of hidden layers are important parameters. These determine the effects of NBM. The evaluation metrics for different combinations of these two parameters are shown in Fig.9. Increasing the number of time steps and hidden layers will lead to a waste of training resources; therefore, the number of time steps is set from 0 to 10, and the number of hidden layers is set from 2 to 14. The results show that fewer time steps can improve the performance of the model when the number of implied layers is extremely small. The model performs best when the time step is greater than 0 and the number of implied layers is greater than 6. The evaluation metrics for both the training and test sets were extremely low. This indicates that the model has a relatively high degree of generalization and accuracy. The final structure with a time step of 6 and 10 implied layers is chosen, at which point the model has the lowest RMSE and MAPE metrics.

Fig. 9
(Color online) The performance of the LSTM-AE model on difference combination. Red box indicates the best combination.
pic

The NBM constructed by the LSTM-AE neural network can reconstruct normal samples well, however, it cannot reconstruct abnormal samples, which is the main reason why the reconstruction error can effectively distinguish normal from abnormal samples.

In this study, the experiment proved the rationality of the reconstruction error. Figures 11 and 12 describe the reconstruction result and error curves under normal and four abnormal conditions. Fig.11 describes the reconstruction result and error curve of the normal samples. Under normal circumstances, the NBM model has high reconstruction accuracy, and the reconstruction error of the sample is below 0.05. However, the change in parameters leads to an increase in the reconstruction error when an anomaly occurs, and the reconstruction errors under the four abnormal conditions are all greater than 0.06. Figure 12 describes in detail the reconstruction accuracy of samples in four common anomalies, compared with Fig.11, the reconstruction accuracy of the sample is reduced. Thus, compared with normal samples, anomaly samples are more difficult to reconstruct using the NBM. The aforementioned experimental results prove that the reconstruction error obtained by the NBM can significantly reflect the difference between normal samples and anomalies. Thus, it is reasonable to treat the reconstruction error as an efficient feature for anomaly detection.

Fig. 11
Example for reconstruction with normal condition. (a) reconstruction precision. (b) reconstruction error.
pic
Fig. 12
Example for reconstruction with four different anomalies. (a) reconstruction precision and error under abnormal current size. (b) reconstruction precision and error under the non-action point anomaly. (c) reconstruction precision and error under the backward movement of the action point. (d) reconstruction precision and error under slow current rise.
pic
4.2
Anomaly detection based on XGBoost

Because the LSTM-AE model training data are normal coil currents, anomalous reconstruction errors were obtained when the validation set containing anomalous data was input to the trained model. Owing to limited anomaly data in the cold CRDM re-qualification test, we set the anomaly data with a small amount of noise to obtain 2400 anomaly reconstruction error data, each of which contained 400 data, where each data contained four cycles. These anomalous reconstruction error data were used as the sample set for XGBoost, with 70% of the sample set as the training set, and 30% as the test set. The reconstruction error obtained by the NBM was input into the XGBoost anomaly detection algorithm as the final feature, and four common anomaly detection results were obtained. Fig. 13 shows an example of detection results for four common anomalies, red points represent the detected outliers. It can be observed that significant outliers can be detected, and contain some early and small anomalies.

Fig. 13
Detection results of XGBoost. Red points represent detected outliers. (a) abnormal current size. (b) the non-action point anomaly. (c) the backward movement of the action point. (d) slow current rise.
pic

This shows that it is reasonable to use XGBoost to detect abnormal reconstruction features. To better evaluate the performance of the anomaly detection algorithm, this study used the most common indicators to measure the performance of anomaly detection. Precision and recall are the two most commonly used evaluation indices for classification purposes. Precision represents the ratio of the correct total number predicted by the model to the total number predicted, and recall rate is the ratio of the correct total number to the actual target number. The correct total number predicted in the classification was the correct number of categories corresponding to the maximum probability. The precision and recall rate formulae are as follows: Precision=TPTP+FP, (21) Recall=TPTP+FN. (22) The true positive (TP) is the number of correct predictions. The false positive (FP)is the number of predicted errors. A false negative (FN) is the number not detected. TP + FP denotes the number of correct predictions. TP + FN is expressed as the actual correct number.

Table 1 shows the accuracy and recall of the four anomaly detection methods. It can be observed that XGBoost has good precision and recall rates for the detection results of these four anomalies. To further validate the performance of the classification models. Figure 14 shows the receiver operating curve (ROC) under the four abnormal conditions characteristic curve. The transverse axis is the false positive rate (FPR) of the model and the longitudinal axis is the true positive rate (TPR). It can be observed that the ROC curves in four abnormal cases are closer to the upper left of the two-dimensional plane. This shows that XGBoost has good classification performance for outlier detection.

Table 1
Detection errors with XGBoost for four different anomalies.
abnormal situation 1 2 3 4
Precision 99.67 99.62 99.56 99.73
Recall 91.58 96.33 96.85 92.32
Show more
Fig. 14
ROC curves of four different anomalies.
pic
4.3
Comparison

Through the aforementioned experiments, the reasonability of the proposed anomaly method is proved. To verify the detection performance of the proposed method, three groups of Comparative experiments were performed in this study. The first group of experiments verifies the prediction effect of the NBM based on the LSTM-AE model. The second group of experiments was designed to verify the performance of the XGBoost anomaly detection.

To demonstrate the superiority of the LSTM-AE method more clearly in establishing the NBM, the proposed method was compared with common reconstruction or prediction models. Comparisons between LSTM-AE, AE, LSTM, and ARIMA models are shown in Fig. 15. AE and LSTM-AE are reconstruction models, whereas ARIMA and LSTM are prediction models. The LSTM model adopts a two-layer structure; the number of neurons in the two layers is 32 and 64, respectively. The parameters p, d, and q of the ARIMA model were 10, 1, and 10, respectively. It can be observed that the AE neural network has the worst effect, the LSTM model has a better prediction effect than the ARIMA model, and the LSTM-AE model has a better reconstruction effect than the other three models. Table 2 lists the RMSE and MAPE of the four methods for the training and test datasets. Compared with the AE method, the LSTM-AE method reduced the RMSE by 96.51% and MAPE by 88.69% on the training dataset. Compared with ARIMA, the LSTM-AE method reduced the RMSE and MAPE by 77.55% and 76.51%, respectively, on the training dataset. Compared with LSTM, the LSTM-AE method reduced the RMSE by 64.51% and MAPE by 56.95% on the training dataset. LSTM-AE was also significantly better than the remaining three methods on the test dataset. Consequently, the LSTM-AE method can better establish the normal behavior model (NBM) of the CRDM coil current.

Table 2
Comparison of error of different forecasting models
Model Train Test
  RMSE MAPE RMSE MAPE
AE 0.315 0.849 0.317 0.855
ARIMA 0.049 0.452 0.049 0.396
LSTM 0.031 0.223 0.036 0.265
LSTM-AE 0.011 0.096 0.009 0.093
Show more
Fig. 15
(Color online) Comparison of forecasting performance between ARIMA, LSTM, AE and LSTM-AE model.
pic

To further verify the performance of XGBoost anomaly detection, the proposed method was compared with common classification algorithms. The original data were input to the common classification algorithm after the reconstruction residual was obtained by the NBM model and compared with XGBoost, such as XGBoost, OCSVM, Iforest, LOF, and KNN. Among them, the OCSVM chooses the radial basis function as the kernel function. The penalty coefficient was set to 10. The insensitivity coefficient was set at 0.28. The number of forest trees was set at 100. The number of neighbors in the LOF and that of the KNN was set to 20.

The ROC curves for the five algorithms are shown in Fig. 16. Compared with OCSVM, Iforest, LOF, and KNN, the ROC curve of XGBoost is closer to the upper left and has a larger area of the two-dimensional plane. This indicates that XGBoost has a better anomaly classification effect.

Fig. 16
Visualization of ROC for different algorithms.
pic

The specific values of the precision and recall indicators of the five algorithms for the four common anomalies are listed in Table 3. In the four abnormal cases, the precision and recall indexes of XGBoost were significantly better than those of SVM, Iforest, LOF, and KNN. This indicates that XGBoost has a better classification accuracy. XGBoost improves the precision by 31.77% and recall by 31.36% on average compared with the other three classification algorithms. This verifies the effectiveness of XGBoost for CRDM coil current anomaly detection and identification.

Table 3
Performance comparison of different classification algorithms
Classification Indicator Abnormal situation
algorithm types (%) 1 2 3 4
KNN Precision 70.42 64.72 63.98 67.63
  Recall 62.33 63.41 64.13 61.47
LOF Precision 75.41 69.22 68.41 69.79
  Recall 64.52 65.34 65.19 66.26
Iforest Precision 80.26 74.10 72.47 78.15
  Recall 72.49 73.66 73.16 70.18
OCSVM Precision 91.25 87.56 86.85 90.14
  Recall 81.56 82.52 83.12 80.41
XGBoost Precision 99.67 99.62 99.56 99.73
  Recall 91.58 96.33 96.85 92.32
Show more
5

CONCLUSION

This study presents a hybrid anomaly detection method based on the combination of the LSTM-AE neural network and XGBoost algorithm for the coil current of CRDM. The proposed method utilizes the LSTM-AE neural network to automatically extract nonlinear and sequential features of the coil currents and builds a normal behavior model to reconstruct the residual under abnormal conditions. The long series and periodic historical current data can be handled and predicted using the LSTM-AE model. The XGBoost classification algorithm was employed to identify outliers in the reconstructed errors and avoid defects. The proposed method inherits the generalization and classification abilities of these two techniques to realize the accurate detection of abnormal points and anomalies for CRDM. The results demonstrate that the proposed method enhances the prediction performance and classification accuracy compared with other detection methods. Finally, it can be confirmed that this solution offers promising applications for improving the safety and reliability of nuclear reactors.

References
[1] J.W. Lee, S.S. Cho, D.O. Kim et al.,

Dynanic characteristics on the CRDM of SMART reactor

. J. Korean Mucl. Soc. 34, 1105-1111 (2010). doi: 10.3795/KSME-A.2010.34.8.1105
Baidu ScholarGoogle Scholar
[2] A. Qluwasegun, J.C. Jung,

The application of machine learning for the prognostics and health management of control element drive system

. Nucl. Eng. Technol. 52, 2262-2273 (2020). doi: 10.1016/j.net.2020.03.028
Baidu ScholarGoogle Scholar
[3] C. Y. Peng, P. He, X. W. Peng et al.,

Research on Methods of Failuwre Analysis for ControlRod Drive Mechanism Based on Structure Moise Detecting Technique

. Nucl. Power Eng. 41, 214-216 (2020). doi: 10.13832/j.jnpe.2020.03.0214
Baidu ScholarGoogle Scholar
[4] X. Y. Xie, H. Peng, Z. W. Zhang et al.,

Study on monitoring of action assembly of control rod drive mechanism in nuclear power reactors

. Nucl. Power Eng. 40, 90-94 (2019). doi: 10.13832/j.jnpe.2019.02.0090
Baidu ScholarGoogle Scholar
[5] P.E. Huang, P. He, S.Q. He et al.,

Study on condition Monitoring and fault diagnosis of control rod driving mechanisn

. Nucl. Electron. Detect. Technol. 3, 33 (2013). doi: 10.3969/j.issn.0258-0934.2013.03.017 (in Chinese)
Baidu ScholarGoogle Scholar
[6] J.G. Zhang, H.J. Yian, Y.Q. Wu et al.,

Research on the electromagnetic structure of movable coil electromagnet drive mechanism for reactor control rod

. J. Nucl. Sci. Technol. 44, 163-170 (2007). doi: 10.1080/18811248.2007.9711270
Baidu ScholarGoogle Scholar
[7] C. Ma, H. L. Bo, S. Y. Jiang et al.,

Study of maximal load capacity of moveable coil electromagnetic drive for reactor control rod

. Nucl. Power Eng. 27, 55-57 (2006). doi: 10.1300/j064v28no110
Baidu ScholarGoogle Scholar
[8] S.T. Ling, W.Q. Li, C.X. Li et al.,

Flow field fusion simulation method based on model features and its application in CRDM

. Nucl. Sci Tech. 33, 32 (2022). doi: 10.1007/s41365-022-01022-0
Baidu ScholarGoogle Scholar
[9] J. L. Chen, P.C. Ai, D. Wang et al.,

FPGA implementation of neural network accelerator for pulse information extraction in high energy physics

. Nucl. Sci. Tech. 31, 46 (2020). doi: 10.1007/s41365-020-00756-z
Baidu ScholarGoogle Scholar
[10] X.K. Ma, H.Q. Huang, Q.C. Wang et al.,

Estimation of Gaussian overlapping nuclear pulse parameters based on a deep learning LSTM model

. Nucl. Sci. Tech. 30, 171 (2019). doi: 10.1007/s41365-019-0691-2
Baidu ScholarGoogle Scholar
[11] A. Stetco, F. Dirnmohannadi, X.Y. Zhao et al.,

Machine learning methods for wind turbine condition monitoring: A review

. Renewable Energy 133, 620-635 (2019). doi: 10.1016/j.renene.2018.10.047
Baidu ScholarGoogle Scholar
[12] T.V. Tung, B.S. Yang,

Machine fault diagnosis and prognosis: The state of the art

. Intern. J. Fluid Mach. System. 2, 61-71 (2009). doi: 10.5293/IJFMS.2009.2.1.061
Baidu ScholarGoogle Scholar
[13] N. Alghanmi, R. Alotaibi, S.M. Buhari et al.,

Machine learning approaches for anomaly detection in IoT: An overview and future research directions

. Wireless Pers Commun. 122, 2309-2304 (2020). doi: 10.1007/s11277-021-08994-z
Baidu ScholarGoogle Scholar
[14] M. H. Gu, J. J. Fei, S. L. Sun et al.,

Online anomaly detection with sparse Gaussian processes

. Neurocomputing. 403, 383-399 (2020). doi: 10.1016/j.neucom.2020.04.077
Baidu ScholarGoogle Scholar
[15] R.K. Pandit, D. Infield,

SCADA based wind turbine anomaly detection using Gaussian Process (GP) models for wind turbine condition monitoring purposes

. IET Renewable Power Generation 12, 1249-1255 (2018). doi: 10.1049/iet-rpg.2018.0156
Baidu ScholarGoogle Scholar
[16] J. Song, H. Takakura,

Unsupervised anomaly detection based on clustering and multiple one-class SVM

. Trans. Inst. Electron. 92, 1981-1990 (2009). doi: 10.1587/transcom.E92.B.1981
Baidu ScholarGoogle Scholar
[17] X. Wang, S.H. Ahn,

Real-time prediction and anomaly detection of electrical load in a residential community

. Appl. Energ. 259, 114145 (2020). doi: 10.1016/j.apenergy.2019.114145
Baidu ScholarGoogle Scholar
[18] H. Sarmadi, A. Karamodin,

A novel anomaly detection method based on adaptive Mahalanobis-squared distance and one-class kNN rule for structural health monitoring under environmental effects

. Mech. Syst. Signal. Proc. 140, 106495 (2020). doi: 10.1016/j.ymssp.2019.106495
Baidu ScholarGoogle Scholar
[19] M. Ding, T. Hui,

PCA-based network traffic anomaly detection

. Tsinghua. Sci. Technol. 21, 500-509 (2016). doi: 10.1109/TST.2016.7590319
Baidu ScholarGoogle Scholar
[20] A. Sheinker, M.B. Moldwin,

Magnetic anomaly detection (MAD) of ferromagnetic pipelines using principal component analysis (PCA)

. Meas. Sci. Technol. 27, 045104 (2016). doi: 10.1088/0957-0233/27/4/045104
Baidu ScholarGoogle Scholar
[21] C. Li, L. B. Guo, H. Guo et al.,

Similarity-measured isolation forest: An anomaly detection method for machine monitoring data

. IEEE Trans. Instrum. Meas. 70, 3512512 (2021). doi: 10.1109/TIM.2021.3062684
Baidu ScholarGoogle Scholar
[22] L. Puggini, S. McLoone,

An enhanced variable selection and isolation forest based methodology for anomaly detection with OES data

. Eng. Appl. Artif. Intel. 67, 126-135 (2018). doi: 10.1016/j.engappai.2017.09.021
Baidu ScholarGoogle Scholar
[23] M. Canizo, I. Triguero, A. Conde et al.,

Multi-head CNN-RNN for multi-time series anomaly detection: An industrial case study

. Neurocomputing 363, 246-260 (2019). doi: 10.1016/j.neucom.2019.07.034
Baidu ScholarGoogle Scholar
[24] Y. Yan, G. Sheng, Y. Chen et al.,

An method for anomaly detection of state information of power equipment based on big data analysis

. Proceeings of the CSEE 35, 52-59 (2015). doi: 10.13334/j.0258-8013.pcsee.2015.01.007 (in Chinese)
Baidu ScholarGoogle Scholar
[25] T. Ergen, S. Kozat,

Unsupervised anomaly detection with LSTM neural networks

. IEEE T. Neur. Net. Lear 31, 3127-3141 (2020). doi: 10.1109/TNNLS.2019.2935975
Baidu ScholarGoogle Scholar
[26] A. Coluccia, D. Alconzo, F. Ricciato,

Distribution-based anomaly detection via generalized likelihood ratio test: A general maximum entropy approach

. Comput. Netw 57, 3446-3462 (2013). doi: 10.1016/j.comnet.2013.07.028
Baidu ScholarGoogle Scholar
[27] D. Zhang, L. Qian, B. Mao et al.,

A data-driven design for fault detection of wind turbines using random forests and XGBoost

. IEEE Access, 21020-21031 (2018). doi: 10.1109/ACCESS.2018.2818678
Baidu ScholarGoogle Scholar
[28] J. Zeng, C. Peng, P. He et al.,

Research on Current Monitoring and Fault Diagnosis Technology for Control Rod Drive Mechanism

. Nucl. Power Eng. 40, 172-175 (2019). doi: 10.13832/j.jnpe.2019.01.0172
Baidu ScholarGoogle Scholar