Introduction
Concept of Digital Twins (DTs)
The concept of digital twins (DTs) has gained significant traction in recent years, particularly in the engineering and industrial disciplines [1]. A DT refers to a virtual model that closely mirrors a real-world physical product, system, or process, serving as an effectively indistinguishable digital counterpart for practical purposes such as simulation, integration, testing, monitoring, and maintenance. DTs have been recognized as the fundamental premise of product lifecycle management, encompassing the entire lifecycle of the physical entity they represent, including creation, construction, operation/support, and disposal [2-6].
The value and fidelity of a DT representation depend on the specific use cases, considering that its granularity and complexity are tailored accordingly. Notably, a DT can be conceptualized and utilized even before its physical counterpart exists, allowing for comprehensive modeling and simulation of the intended entity’s lifecycle.
The origin of the DT concept can be traced back to NASA, which pioneered its practical definition in 2010 as part of its efforts to enhance the simulation of spacecraft using physical model representations [2, 6]. Since then, the development of DTs has progressed in tandem with advances in product design and engineering. Traditional manual drafting techniques have given way to computer-aided drafting and design and model-based system engineering approaches, which establish a direct link between the DT and signals from its physical counterparts.
In recent years, DTs have garnered significant attention in various industrial sectors. Notably, their application in the chemical processing industry has demonstrated versatility [7]. In addition, DTs are increasingly utilized in simulation-based vehicle certification and fleet management [8], as well as in addressing the challenges of unpredictable and undesirable behaviors in complex systems [9, 10]. A specific area of growth has been observed in the nuclear plant industry, where digital twin architectures have been developed for the management and monitoring of nuclear plants, reflecting the expanding scope and potential of technology [11, 12].
Previous Work on Reactor Operation Digital Twins (RODTs)
A significant amount of research has been conducted to explore the potential applications of reactor operation digital twins (RODT) in the nuclear energy domain. RODTs represent a specific instantiation of the DT concept, focusing on the numerical representation of nuclear reactors for real-time prediction, optimization, monitoring, control, and decision-making during their operational stage [13].
A previous study [13] introduced the first prototype of an RODT specifically tailored for the prediction of neutron flux and power fields in the operational stage of the HPR1000 reactor core [14]. The prototype demonstrated the feasibility of using the RODT for online monitoring and prediction of key reactor parameters, thereby enhancing operational understanding and performance assessment.
In [15], the forward model of the RODT was realized through the application of a nonintrusive reduced-order model constructed using an SVD autoencoder to learn the nonlinearity of the field distribution. Machine learning techniques, specifically leveraging k-nearest neighbour (KNN) and decision trees, are employed to establish forward mapping. Additionally, the inverse model adopts a generalized latent assimilation method for accurate estimation of the model parameters.
While demonstrating promise, opportunities remain for improving the efficacy of RODT for practical deployment. In particular, the inverse solver relies on a computationally expensive methodology that limits its potential integration within an online framework. Therefore, the aim of this work [16] is to enhance the key capabilities of the RODT. This includes proposing an advanced differential evolution algorithm to upgrade the inverse solver for improved efficiency and accuracy of parameter estimation. In addition, uncertainty quantification was conducted to validate the RODT’s performance considering noisy observational data. Numerical validation experiments were performed across representative domains for an HPR1000 pressurized water reactor core to demonstrate its potential for engineering applications.
The key steps for constructing an RODT include: (i) training a reduced-order model (ROM) using model order reduction (MOR) [17-20] techniques as well as an efficient forward model and (ii) adaptation and implementation of inverse models to infer the input parameters and the resulting field distribution.
Introduction to Reduced Order Models (ROMs)
Many mathematical models encountered in real-life processes present computational challenges when employed in numerical simulations because of their inherent complexity and large dimensionality. To address these challenges, MOR techniques have been developed to reduce the computational complexity of these problems, particularly in the context of simulations involving large-scale dynamic and control systems. The ROM was computed as an approximation of the original model by reducing the dimensionality or degrees of freedom associated with the model.
The preparation of an ROM model involves two distinct phases, the offline and online phases, which are integral to the creation of a simplified representation of a high-fidelity full-order numerical model, enabling efficient real-time simulations and analysis. The offline phase focuses on determining the inherent low-dimensional structure of the underlying full-order model and deriving the reduced basis functions from high-fidelity snapshots of the system. Intrusive MOR techniques [21] are employed when a comprehensive understanding of the governing equations and numerical strategies employed in the full-order model is available. This requires access to the detailed numerical framework of the full-order model, which may involve proprietary commercial codes. By projecting a full-order model onto a reduced space, intrusive ROMs provide a concise representation with reduced computational complexity.
However, in practical engineering scenarios [22-24], detailed information regarding a full-order model is often unattainable, and the solvers or codes implementing these models are treated as black-box entities. This limitation poses a challenge to the implementation of traditional intrusive MOR methods. To address this issue, nonintrusive MOR techniques have been developed. Non-intrusive MOR approaches aim to construct ROMs without relying on detailed knowledge of the numerical framework of a full-order model. Instead, these methods establish an input-output mapping between the input parameters and the reduced basis through data-driven techniques such as interpolation, regression, or machine learning, approximating the behavior of the full-order model based on a reduced set of parameters and basis functions.
ROMs prepared through the offline and online phases play a critical role in the development of RODTs. RODTs require the integration of high-fidelity models and solvers with real-time capabilities, and ROM-based emulators provide promising solutions [21-23, 25]. Leveraging ROM-based forward solvers, RODTs can perform both one-to-one forward simulations and one-to-many simulations for inverse problems, encompassing parameter identification [26-29], data assimilation [24, 30-35], sensitivity/uncertainty quantification [36-40], and optimization/control [37, 41-47].
Forward Model Construction
In line with the framework RODT, as investigated in a previous study [13, 15], we selected KNN [48] to construct the forward model. The choice of KNN is justified by its numerous advantages, which render it a suitable candidate for the forward model in this context.
First, it is a nonparametric algorithm, meaning that it does not assume any specific functional form for the relationship between the inputs and outputs. Second, KNN is relatively simple to implement and computes efficiently, particularly for low-dimensional input problems. The complexity of the algorithm depends primarily on the number of training samples, making it suitable for problems with moderate dataset sizes [49]. In a recent study [50], an investigation was conducted to explore the possibility of utilizing different values of K for each dataset, considering the information provided by the correlation matrix.
Inverse Model Construction
Generally, the inverse problem can be divided into two categories:
• Parameter Identification. Solving the parameter vector based on the information from observation vectors: In this scenario, we aim to estimate the unknown parameters of the system based on the available observation vectors. The observation vectors contain information about the system’s behavior or response under conditions that are characterized by the parameter vector.
• State Estimation. Solving the field distribution based on the information from observation vectors: In this scenario, the objective is to determine the unknown field distribution of the nuclear system based on the available observation vectors.
Contribution of this Work
A previous work [13, 15, 16] demonstrated the feasibility of employing the RODT framework in engineering applications. However, a need exists for a systematic and in-depth investigation of the core components of RODT, namely parameter identification and state estimation. In addition, the development of mature and modular software is crucial to facilitating the application of RODT in practical scenarios. This motivated us to develop an efficient and appropriate surrogate inverse model to accurately predict the parameters and states. This also motivated our research on hybrid optimization approaches that can effectively balance coarse- and fine-grid search (coarse–fine-grid search) methods, which are discussed in the following sections.
The contributions of this research are as follows:
• Investigation and implementation of novel gradient-free optimization algorithms specifically tailored to handle the challenges posed by the discontinuous surrogate forward mapping constructed with the KNN forward model of our RODT and comparing the convergence profile, stability, and accuracy-time performance.
• Exploration and evaluation of the proposed coarse–fine-grid search methods within various global optimization approaches, aiming to strike a balance between accuracy and computational efficiency in parameter identification and state estimation.
• Development and deployment of efficient modular RODT software integrated with the aforementioned algorithms into the inverse problem solver, which applies to other sustainable energy systems.
These contributions focus on advancing state-of-the-art RODT by addressing the unique challenges encountered in solving inverse problems in nuclear engineering and proposing novel methodologies for improved parameter identification and state estimation. Furthermore, our work addresses the critical challenge of achieving real-time, high-precision simulations in nuclear energy DT applications, significantly enhancing the capabilities of DTs in the nuclear energy sector and facilitating improved monitoring, control, and optimization of nuclear energy systems.
Methodology
The modeling process of an RODT is classified into offline and online stages.
• Offline phase: training a non-intrusive ROM, which includes (i) preparation of the full order snapshots; (ii) preparation of the reduced basis; (iii) preparation of the coefficient; and (iv) preparation of the surrogate forward model to predict the coefficient with the information of the parameter.
• Online phase: parameter identification and state estimation, which includes (i) building an inverse model to predict the parameter based on the information of the observation data (clean or with observation noise) and (ii) building an inverse model to predict the coefficient based on the information of the observation data (clean or with observation noise).
The methodology is illustrated in Fig. 1.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F001.jpg)
Offline phase: Training a Non-Intrusive ROM
The variation in the power field Φ in a nuclear core is characterized by physical laws that are generally described implicitly by a governing equation [51], such as the neutron transport or diffusion equations, which is written as Equation 1.
Preparation of the Full-Order Snapshots
In this study, the numerical solution of Eq. 1 was obtained using the proprietary CORCA-3D code package developed at the Nuclear Power Institute of China (NPIC) [53]. The CORCA-3D code is treated as a black-box solver; it can be substituted with other proprietary codes as needed.
To construct a non-intrusive ROM, it is necessary to gather a set of full-order solution snapshots. We sample μi in the parameter domain 𝒟 to obtain the discrete set
The knowledge of this manifold is implicitly dependent on the knowledge of the parameters μ, and for better construction of a non-intrusive model, we endeavor to find a low-rank representation for the solution snapshot set 𝑴. Reduced-basis methods (RB) [54-56] indicates that in state where Φ is sufficiently regular, it can be approximated by an n-dimensional reduced basis {qi}, as follows:
Preparation of the Reduced Basis
For simplicity and time efficiency, we chose the singular value decomposition (SVD) method for generating the reduced basis. The steps are as follows:
Correlation Matrix. The correlation matrix C is computed by taking the inner product of each pair of snapshots, as expressed by the following equation:
Proper orthogonal decomposition (POD). The jth POD basis vector qj, which is independent of parameter μ, is obtained as a linear combination of snapshots.
Collection of the best basis minimizing the Frobenious error. The solution snapshot set is denoted as
Preparation of the Coefficient
Following the aforementioned procedure, the coefficient is constructed as the projection of the snapshot matrix S onto the vector space spanned by the POD basis
For any
Moreover, in matrix form, the coefficient matrix
Training a Forward Model
The goal of the forward model is to construct an efficient surrogate mapping model between the parameter and coefficient, as follows:
The response of the ROM model, denoted as
We design the cost function as Eq. 12 based on the residual. The forward model is used in the search of the optimal approximation for the operator ϕ in the functional set
Previous research [13] evaluated various forward model approaches, among which the KNN algorithm demonstrated exceptional accuracy and efficiency. In this study, we specifically selected the KNN algorithm [48] as the forward model to exploit its full potential. To enhance the performance of the KNN algorithm, we focused on two crucial parameters that significantly influence its efficacy.
The first parameter relates to the voting distance in KNN, wherein we investigate the impact of utilizing either the Euclidean distance or Manhattan distance as the metric for determining the nearest parameter choice in the training set. Notably, these two distances can be expressed as Minkowski norms with an exponent of p = 1, 2, corresponding to the Euclidean and Manhattan distances, respectively. The Minkowski distance, a measure of the discrepancy between the testing and training parameters, can be mathematically represented as follows:
Another vital parameter is K, which refers to the number of candidates voted for. The optimal choice of K can be determined by plotting the relative L2 error of the predicted observation and true vectors, which is expressed as follows.
From Fig. 2, we can deduce that when K = 5, our forward model using KNN performs with the best prediction accuracy, and the variation in the error is relatively smooth, which corresponds to the continuous variation in the value of the field at the corresponding position of the field manifold snapshot.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F002.jpg)
In contrast, the stair-shaped observation data predictions in Fig. 3 demonstrate that the KNN forward surrogate model has a step-like progression as neighbors are adjusted between discrete datasets and thus are not continuous or differentiable. In this manner, well-developed state-of-the-art algorithms based on gradient information for solving optimization and inverse problems did not work well in our case. To meet this demand, we investigated other novel approaches, which are discussed in detail in the following sections.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F003.jpg)
Online phase: Parameter Identification and State Estimation
The inverse problem in nuclear engineering refers to the task of determining the model parameters that yield the observed data, as opposed to the forward problem, which involves predicting the data based on the given model parameters. The objective is to determine the model parameters μ* that approximately satisfy Equation 12.
The nature of the inverse problem depends on whether operator ϕ is linear or nonlinear. In most cases, particularly when solving systems governed by the neutron flux diffusion function, the inverse problem is nonlinear, owing to the nonlinearity of the forward map. Thus, we attempt to develop nonlinear algorithms, gradient-free algorithms, or algorithms based on the initial guess given by fast linear algorithms, such as KNN.
Setup of Inverse Model and Optimization
In the context of optimal control theory, the governing equations that describe the behavior of a physical system are commonly referred to as state equations. However, in many practical scenarios, interest lies not only in the physical state itself but also in its impact on certain objects or quantities. Furthermore, only a limited amount of data can often be obtained from the physical state. To address these considerations, an additional operator denoted by
Comparison of Different Approaches for Solving Inverse Models
To solve inverse problems, machine learning and optimization methods can be employed to search for the optimal parameters or field distributions that best fit the observed data [58]. Additionally, as mentioned in Sect. 2.1.4, when we employ the KNN algorithm to construct the surrogate forward model, the surrogate forward mapping is non-differentiable and discontinuous. This poses a challenge because the traditional approaches used to solve continuous optimization and inverse problems [59] are not applicable in this context. This necessitates the exploration of novel inverse problem-solving methods suitable for our particular problem. Hence, this study investigates various deterministic optimization algorithms, metaheuristic algorithms, and their hybrids to address this challenge.
Exhaustive Direct Search with Latin Hypercube Sampling (LHS)
To evaluate and compare the different approaches to solving inverse models, it is essential to establish a benchmark for measuring their performance. In this section, we propose using exhaustive direct search (EDS) with latin hypercube sampling (LHS) as the benchmark method for inverse problems. It combines the accuracy of EDS methods, which have been widely used since the 1960s [60], with the sampling efficiency of LHS.
LHS [61] is a sampling technique that ensures a comprehensive exploration of the solution space while maintaining desirable properties. The range of each variable was divided into equally probable intervals, and the sample points were strategically placed to satisfy the requirements of a Latin hypercube, where each sample was the only one in its corresponding axis-aligned hyperplane. The advantage of LHS lies in its independence from the number of dimensions, which allows for efficient sampling without the need for an increasing number of samples. Furthermore, LHS facilitates sequential sampling, enabling the tracking of the already selected samples.
Exhaustive Direct Search with 2 Steps Latin Hypercube Sampling (LHS2STEPS)
To enhance the benchmark algorithm and mitigate the uncertainties introduced by the randomness of the sampling process, [13] introduced a two-step sampling strategy. This strategy involves further sampling the neighborhoods of the n1 best candidates generated during the initial stages of the LHS algorithm. For the detailed pseudocode, please refer to 3.3, excluding the part related to ALHS.
Exhaustive Direct Search with Assembled 2 Steps Latin Hypercube Sampling (ALHS)
To improve our benchmark algorithm and further reduce the uncertainty caused by the randomness of the sampling process, we introduce a mean mechanism at the end of the classical LHS algorithm to calculate the mean of the n1 top-rated candidates in the five clones, which are spread around the n2 top-rated candidates in the first sampling process. The concept is illustrated in Algorithm 1.
Inputs: | |||
Observations: Yo | |||
Initial guess: μinitial | |||
Forward function: FML | |||
Transformation operator: H | |||
First stage sampling number: ns, 1 | |||
Second stage sampling number: ns, 2 | |||
for i from 2 to n1 do | |||
▷ | n1=5 in this article. | ||
end for | |||
for j from 1 to n1 do | |||
end for | |||
for k from 1 to n2 do | ▷ | Start of the main part of ALHS, n2=10 in this article. | |
end for | ▷ | End of the main part of ALHS. | |
Outputs: | |||
Parameter: |
|||
Reconstructed field: |
Efficient Deterministic Machine Learning Algorithm: KNN
In addition to the benchmark LHS method, another efficient deterministic algorithm that can be considered for solving inverse models is the KNN algorithm. KNN is a nonparametric classification and regression algorithm that can be adapted to solve inverse problems.
The KNN algorithm operates based on the principle of similarity. Given a new input data point, KNN finds the K-nearest neighbors in the training dataset based on the Euclidean distance and then generates the output value by averaging the output values of the K-nearest neighbors. The selection of the optimal value of K can be determined by graphing the relative L2 error between the predicted observation vector and the true vector, as shown in Fig. 2. The analysis in Fig. 2 reveals that selecting K = 1 yields the optimal outcome in terms of minimizing the L2 prediction error, thereby emphasizing its advantageous accuracy.
Metaheuristics Algorithms with Physical Constraints
Recently, the application of metaheuristic algorithms with physical constraints [62-64] has gained significant attention for inverse modeling. These algorithms offer powerful gradient-free optimization techniques that can effectively handle complex inverse problems while satisfying the physical constraints imposed by system dynamics.
In this manner, we implemented and adapted practical algorithms and constructed related solvers that were integrated into the developed RODT framework. These solvers are designed to address the inverse problem presented by Equation 12.
Differential Evolution Algorithm (DE)
The differential evolution (DE) algorithm [65] offers a promising approach for tackling inverse modeling problems. By leveraging its ability to handle nonlinear and nonconvex objective functions, DE can effectively explore the parameter space. Furthermore, DE provides mechanisms to incorporate physical constraints, ensuring that the generated solutions adhere to system dynamics. The concrete procedure of the DE algorithm can be found in a previous work [13], in which we used the best mutation strategy recommended in [66] for the mutation step in DE, and the mutation donor
Particle Swarm Optimization Algorithm (PSO)
Particle swarm optimization (PSO) [67] is a population-based optimization technique inspired by the collective behavior of bird flocking or fish schooling. It involves particles representing potential solutions that update their positions based on the personal and global best positions. The algorithm iterates a set number of times by adjusting the particle velocities and positions. The best solution, represented by the global best position with the lowest objective function value, is returned.
Fast Simulated Annealing Algorithm (FSA)
The simulated annealing (SA) [68] algorithm, which was initially proposed for solving the well-known travelling salesman problem (TSP) [69], is a popular approach for solving optimization problems in various fields [70], including nuclear engineering [71].
SA was inspired by metallurgical annealing, in which a metal is heated and gradually cooled to achieve a more ordered configuration. Similarly, the SA starts from a high-energy state (initial solution) and progressively lowers the temperature until it converges to a state of minimal energy (optimal solution).
The fast simulated annealing (FSA) [72] is an adaptation of classical simulated annealing (CSA) [68], which improves the computational efficiency and convergence speed. CSA utilizes a local sampling approach and controls the fluctuation variance through an artificial cooling temperature, denoted as Tc(t). FSA modifies the cooling schedule to accelerate convergence using a cooling temperature Tf(t) that decreases reciprocally rather than logarithmically with time t.
Cuckoo Search Algorithm (CS)
In the cuckoo search (CS) algorithm [73], nests are used to represent potential solutions, and eggs within the nests symbolize candidate solutions. Through an iterative process, the algorithm generates new solutions by employing Lévy flights and subsequently evaluates their fitness. If a newly generated solution is superior to an existing one within a given nest, it replaces the incumbent solution. To maintain population diversity and improve the overall quality of solutions, the population is sorted based on fitness, and a predetermined number of solutions are replaced with new solutions generated via random walks. The algorithm terminates upon reaching a predefined maximum number of generations, and the best solution, characterized by the lowest objective function value, is returned as the final outcome.
CS utilizes a balanced combination of a local random walk and a global explorative random walk, controlled by a switching parameter pa. A local random walk is defined as follows:
Artificial Neural Network (NN)
The neural network (NN) algorithm [74], inspired by the behavior of the human brain, simulates interconnected artificial neurons to process information, learn from data, and estimate unknown parameters or states. They consist of interconnected artificial neurons organized into layers that receive input signals, process them through weighted connections, and produce output signals using activation functions. NNs can learn from input-output examples by adjusting the connection weights during training, enabling them to capture underlying patterns and correlations, making them suited for solving physical problems [75].
The combination of the Adam and L-BFGS optimizers in an alternating manner has demonstrated efficacy in training NNs for complex problem solutions [76]. This approach leverages the strengths of the two optimizers to enhance training and improve the overall network performance.
However, owing to NN’s continuous property, it cannot handle discrete problems well, particularly when the test data are noisy, which is further discussed in Sect. 5.
Hybrids of Optimization Methods for Inverse Problem: Balancing Coarse–Fine-Grid Search
To address the challenges posed by the discontinuity of KNN surrogate forward mapping and noisy observations, we propose hybrid algorithms that combine the strengths of deterministic machine-learning methods for coarse-grid search and metaheuristic algorithms for fine-grid search. This methodology can also be found in fields that encompass feature extraction of aerial data [77], automatic free-form optics design [78], and hyperparameter optimization [79]. The objective of a coarse–fine-grid search is to strike a balance between global exploration and local exploitation [80] to achieve accurate, efficient, and less sensitive parameter estimation. The hybrid optimization hub designed accordingly can be further illustrated in Fig. 4.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F004.jpg)
Cuckoo Search with Differential Evolution (CSDE)
In the later stages of the standard CS process, notable issues arise in the form of information waste and slow convergence. These issues stem from the execution of independent evaluations and the lack of sufficient mechanisms for information sharing within the population. Consequently, valuable information fails to propagate effectively, resulting in suboptimal convergence rates and reduced search efficiency.
From an alternative perspective, the underlying biological nature of mutations within the standard DE algorithm was explored. In this context, a novel mathematical elucidation of these mutations is sought. Notably, researchers in the natural sciences have observed that mutations can be understood as sequences of transformations that transpire within a DNA molecule. Furthermore, the transformations pertaining to modifications exhibited a statistical distribution that adhered to the Levy distribution. This finding suggests an analogy between mutations and Levy flight [81]. Drawing inspiration from this observation, it is conceivable to integrate the Levy flight process into the generation of the DE population.
In this regard, we developed the cuckoo search differential evolution (CSDE) algorithm, which lies within the framework of the CS algorithm, integrating the DE mutation operation into the production of the population for each generation. The first step of CSDE is the initialization: Define the objective function
Step 1: Population initialization | ||||
Initialize the population P and the initial position using a random distribution within the search domain range [μlb, μub] | ||||
while |
||||
for each individual μi in population PG do | ||||
Step 2: Population updating | ||||
Generate random indices r1, r2, r3 (different from i) do | ||||
for each dimension j in problem dimension dim(μ) do | ||||
if random(0,1) < CR then | ||||
else | ||||
end if | ||||
end for | ||||
Step 3 Mutation Step | ||||
Generate random indices r3, r4, r5 (different from i) | ||||
Generate a donor vector |
||||
Step 4 Crossover Step | ||||
Generate a trial vector |
||||
for j=1 to dim(μ) do | ||||
if |
||||
else | ||||
end if | ||||
end for | ||||
Step 5 Selection Step | ||||
Evaluate the trial vector |
||||
if |
||||
else | ||||
end if | ||||
Step 6 Abandon nests | ||||
Sort the population PG based on the fitness values in ascending order. | ||||
Determine the number of solutions to be replaced, |
||||
for i = 1 to |
||||
Generate a new solution X in the search space with random walk given in Equation 3.15. | ||||
Evaluate the objective function f for X. | ||||
if f(X) < f(worst solution in PG) then | ||||
Replace the worst solution in PG with X. | ||||
end if | ||||
end for | ||||
end for | ||||
G = G +1 | ||||
end while |
Advanced CSDE (ACSDE) and Advanced DE (ADE)
We utilize the SVC algorithm to classify the test parameters and denote the center of the classified hypercube as the initial machine learning estimate μSVC; then, we use the CSDE and DE algorithms to further explore the parameter space with the initial population spread around the center μSVC. In the SVC process, according to a previous research [15] on the importance rate of the input parameters, the last parameter, temperature Tin, contributes slightly, and we also consider that Tin varies at a relatively small boundary; thus, we only administer the SVC process to the first three parameters. We chose two representative metrics to evaluate the performance of the SVC model.
Confusion Matrix. The confusion matrix [82] provides a comprehensive overview of the binary classification results by displaying the number of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) predictions. This allowed us to assess the accuracy of the model’s by differentiating classes and identifying potential sources of misclassification.
We can infer from Fig. 5 that our SVC classification model predicts the first two parameters with relatively high accuracy while exhibiting confusion concerning the third parameter.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F005.jpg)
F1-score. The F1-score [83], defined in Eq. 3.18, is a popular performance metric for binary classifier models. In Eq. 3.18, TP, FN, FP are the numbers of true positives, false negatives, and false positives classified by the model, respectively.
Based on the Confusion Matrix shown in Fig. 5 and
K-Nearest Neighbour plus Exhaustive Direct Search with Latin Hypercube Sampling (KNNLHS)
Integrating KNN with LHS provides a robust methodology to address inverse problems. To enhance the comprehensiveness of our research, we standardized the grid size for a coarse–fine grid search. This standardization is consistent with the specifications listed in Table 1.
Parameters | St | Bu | Pw |
---|---|---|---|
9.98E-01 | 9.93E-01 | 5.18E-01 | |
Size coarse-grid (SVC) | 41 steps | 250 MWd/tU | 8% FP |
Size fine-grid (DE/CSDE) | 41 steps | 250 MWd/tU | 20% FP |
Relative size coarse-grid (SVC) | 6.67% | 10.00% | 10.00% |
Relative size fine-grid (DE/CSDE) | 6.67% | 10.00% | 25.00% |
Parameter boundary | [0, 615] | [0, 2500] | [20%, 100%] |
In conclusion, the k-nearest neighbour plus exhaustive direct search with latin hypercube sampling (KNNLHS) method is a robust and efficient approach compared to similar strategies. This technique combines the exploratory capabilities of KNN in conducting coarse-grid searches with the efficiency and diversity of LHS in sampling local solution spaces. Consequently, the KNNLHS facilitates precise and efficient parameter estimation, accurate navigation of complex objective functions, and the management of noisy datasets. The effectiveness of this approach and its comparative advantages are discussed in Sect. 5.
Application to Nuclear Reactor
Setup of Reactor Physics Operation Problems
Setup of background information for HPR1000. The objective of constructing the RODT was to accurately predict the power distribution within the HPR1000 reactor core during operation. The core consists of 177 vertical nuclear fuel assemblies, 44 of which are equipped with self-powered neutron detectors (SPNDs) to measure neutronic activity and power fields. Figure 6 provides a visualization of a horizontal slice of the HPR1000 core and an axial slice of an SPND-equipped assembly, with only one-quarter of the core displayed owing to the symmetry along the x and y axes. The gray fuel assemblies represent those containing control rods, whereas the assemblies marked with D indicate the presence of SPNDs. For more detailed information regarding the HPR1000 reactor and generic neutronic physical model, please refer to [14]. Further descriptions of the model for data assimilation and the initial work on RODT can be found in [35] and [13], respectively.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F006.jpg)
During the normal operation of the HPR1000 reactor, two types of control rods were utilized for regulation. The first type, known as compensation rods, is responsible for coarse control and a substantial reactivity reduction. These compensation rods comprise four subtypes: G1, G2, N1, and N2. The second type, called regulating rods (R rods), are employed for fine adjustments to maintain the desired power or temperature [84]. Power evolution within the reactor is influenced by various factors, including the movement of the control rods, the burnup of nuclear fuel, variations in the power level of the reactor core, and fluctuations in the inlet coolant temperature. This evolution was mathematically modeled using two-group diffusion equations [13] and numerically solved using the CORCA-3D code package [85]. Developed at NPIC, the CORCA-3D code is capable of solving 3D few-group diffusion equations, considering thermal-hydraulic feedback, and performing pin-by-pin power reconstruction.
Setup of parameter domain. In this study, we consider CORCA-3D to be an opaque solver in which we input the parameter vector μ and obtain the output Φ. In particular, the power field Φ is constrained to be dependent on a set of "general" parameters that indicate the stage of the reactor’s life cycle:
• Bu: The average burnup of the fuel in the entire core indicates the amount of energy extracted from the fuel and increases over time. Its value ranges between 0 (at the beginning of the fuel’s life cycle) and Bu, max, which is set to 2500 MWd/tU (the end of the fuel’s life cycle). The specific evolution of Bu depends on the reactor’s operational history.
• Pw: The power level of the reactor core ranges from 0.3 to 1 FP (full power).
• Tin: The temperature of the core coolant at the inlet falls within the range of 290 to 300°C.
In this manner, Φ can be implicitly represented by
Setup of observation data and modeling the observation noise. In the context of RODT, the observations denoted by Yo are used to infer the input parameter vector μ and subsequently determine the resulting power field Φ. The arrangement of these observations, organized node by node, is shown in Fig. 6. Each observation value represents the average value over the corresponding node in which the SPNDs are located. It is important to clarify that the observations used in the analysis were not obtained from direct engineering measurements but were derived from numerical simulations of Φ using CORCA-3D.
The noise associated with the observations can be modeled as either a Gaussian or a uniform distribution, depending on the specific requirements. Specifically, the noise term can be mathematically represented as follows:
Considering the prevalence of Gaussian noise in real-world nuclear reactors, our subsequent analysis primarily focused on evaluating the performance of the algorithm under Gaussian noise conditions.
Software implementation We integrated our software into the RODT platform using the following modules:
• rodtPro: This module implements data cleaning and preprocessing techniques based on domain expertise.
• rodtROM: This module constructs ROM through techniques like POD, SVD, and RB.
• rodtML: This module focuses on surrogate modeling and inverse model optimization. It also includes hyperparameter optimization techniques.
• rodtPost: This is a post-evaluation module that defines convergence metrics and conducts comparative analyses using evaluation methodologies derived from peer-reviewed literature.
• rodtUI: This module is responsible for visualizing the results, providing an intrinsic review of the methods, showcasing the optimized hyperparameters, and designing user-specific visualizations.
Practice Metrics for Inverse Model Evaluation
To evaluate the performance of the algorithms, we designed a series of numerical experiments using nuclear reactor data produced by an intrusive program with added noise. We carefully selected metrics related to accuracy and finally defined the normalized reconstruction fields L2 error (E2) and L∞ error (E∞) to evaluate the performance of the comparative algorithms.
The reconstructed field L2 error is defined in Eq. 4.21 to evaluate the average estimation of the field, where
Numerical Results
In this section, we describe the experiments conducted to evaluate the performance of parameter identification and state estimation for the proposed algorithms. The evaluation was performed on a test dataset consisting of 40 specimens from HPR1000, as described in Sect. 4. All the proposed inverse modeling approaches are utilized in the experiments, and it is important to mention that the cost function employed remains consistent with Eq. 2.12. Notably, we focused solely on testing the CSDE method and did not include the CS method. Furthermore, we introduce the proposed accuracy metric, illustrated in Sect. 4.2, to assess its performance.
Accuracy and Stability Analysis of the Parameter Identification Phase
In this section, we describe the tests conducted to evaluate the accuracy of parameter identification. Initially, we assessed the effectiveness of the proposed hybrid methodology, the coarse–fine–grid search, as discussed in Sect. 3.6, using the KNNLHS method as an illustrative example. To visualize this process, we generated scatter plots depicting the predicted values of the parameters St and Pw, which were identified as the two dominant parameters in a previous study [15] (see Fig. 7 for six cases with different input parameter settings.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F007.jpg)
Figure 7 shows a visualization of the local search space in the inverse algorithms. The figures in the first row depict scenarios with varying Pw values ranging from relatively low to moderate to relatively high. Correspondingly, the figures in the second row illustrate cases with varying St values, again encompassing relatively low, moderate, and high values. These figures effectively show the considerable improvements achieved using the KNNLHS method, particularly in the simultaneous optimization of both Pw and St. The visualizations underscore the effectiveness of the method in yielding enhanced optimization outcomes.
The parameter prediction deviation is demonstrated in Fig. 8 over the test set. Owing to space limitations, we chose only KNN, KNNLHS, and FSA for illustration and tested clean and noisy (σ= 5%) observation data. Using the LHS algorithm, the KNN prediction for parameter identification became more accurate, and the prediction deviation for the FSA algorithm was concentrated at relatively low intervals, thus proving its excellent accuracy. In general, the accuracy of burnup, the power level, and the control-rod step are acceptable from an engineering perspective.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F008.jpg)
Comparison of the Convergence Rate among Metaheuristic Algorithms
The convergence rate is a crucial criterion for evaluating algorithms that are integrated with the generation design; thus, in this section, we compare different metaheuristic algorithms and their hybrids. Their convergence performance in 50 iterations (generation) with clean and noisy (σ=1% and σ=5%) observation data is shown in Fig. 9.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F009.jpg)
When the observation data were clean, the FSA exhibited the fastest convergence rate, followed by the advanced differential evolution (ADE). Moreover, when integrated within an SVC coarse-grid search, both the DE and CSDE algorithms demonstrated the ability to escape the local optima in the early stages while incurring a relatively low-cost function, thereby surpassing their standalone performance.
Among the two population-based algorithms, DE outperformed CSDE. This can be attributed to two factors: (i) the adoption of the best mutation method for DE, where we use the optimal mutation strategy for DE while using the classical strategy; and (ii) the relatively straightforward nature of our problem, wherein the inverse problem lacks multiple optima. Consequently, the CS step in each generation of CSDE may lead to unnecessary exploration of the parameter space that is distant from the optimal solution.
In the presence of noise-contaminated observation data, FSA continues to exhibit superior convergence capabilities. However, its advantage diminishes as the level of Gaussian noise reaches 5%, eventually leading to a decline in advantage. Notably, the ADE and advanced cuckoo search differential evolution (ACSDE) failed to deliver satisfactory results under noisy conditions. This is primarily due to the erroneous predictions of SVC, which further leads to inadequate initial estimates provided by population-based algorithms.
Evaluation of State Estimation Phase by Accuracy
In this section, we evaluate various methods in the state-estimation phase based on accuracy. First, we compared the accuracy and stability under two metrics: the relative error in the L2 norm defined by Eq. 21 and the L∞ norm defined by Eq. 22 for field Φ, which is presented in Table 2. The ‘Mean’ values in the tables represent the average accuracy across 40 tested specimens, while the ‘STD’ values indicate the standard deviation of errors over the same set of specimens. The optimal values are indicated in bold.
noise level | 0 | 0.01 | 0.02 | 0.03 | 0.04 | 0.05 |
---|---|---|---|---|---|---|
Mean L2 | ||||||
KNN | 1.0662E-02 | 1.1128E-02 | 1.1461E-02 | 1.1461E-02 | 1.3296E-02 | 1.3541E-02 |
LHS | 4.0457E-02 | 4.0210E-02 | 4.0823E-02 | 4.0754E-02 | 4.1138E-02 | 4.1790E-02 |
ALHS | 3.4294E-02 | 3.4510E-02 | 3.4837E-02 | 3.4865E-02 | 3.5218E-02 | 3.7221E-02 |
DE | 5.3219E-03 | 6.6446E-03 | 7.8485E-03 | 9.2407E-03 | 1.1051E-02 | 1.3777E-02 |
ADE | 3.8119E-03 | 6.8077E-03 | 7.8185E-03 | 9.1585E-03 | 1.1134E-02 | 1.3526E-02 |
CSDE | 1.7029E-02 | 1.6581E-02 | 1.7011E-02 | 1.7390E-02 | 1.8874E-02 | 2.0011E-02 |
ACSDE | 7.6471E-03 | 4.4970E-02 | 4.5157E-02 | 4.6044E-02 | 4.7066E-02 | 4.8563E-02 |
PSO | 1.0093E-02 | 1.1502E-02 | 1.3355E-02 | 1.4640E-02 | 1.6516E-02 | 1.7939E-02 |
FSA | 3.8796E-03 | 5.1561E-03 | 6.5906E-03 | 8.1864E-03 | 1.0106E-02 | 1.2991E-02 |
NN | 2.6737E-02 | 9.8060E-02 | 1.3113E-01 | 1.6029E-01 | 1.8785E-01 | 2.1606E-01 |
LHS2STEPS | 3.1108E-02 | 3.1410E-02 | 3.1664E-02 | 3.2919E-02 | 3.3712E-02 | 3.4975E-02 |
KNNLHS | 7.0615E-03 | 9.0129E-03 | 9.6693E-03 | 1.0514E-02 | 1.2406E-02 | 1.3408E-02 |
STD L2 | ||||||
KNN | 5.3359E-03 | 5.5301E-03 | 5.2201E-03 | 5.2201E-03 | 5.5587E-03 | 5.7721E-03 |
LHS | 2.5494E-02 | 2.5404E-02 | 2.5938E-02 | 2.5916E-02 | 2.5778E-02 | 2.5591E-02 |
ALHS | 2.3719E-02 | 2.3621E-02 | 2.3377E-02 | 2.3131E-02 | 2.2688E-02 | 2.2074E-02 |
DE | 5.3195E-03 | 4.6991E-03 | 4.6245E-03 | 4.6625E-03 | 5.7638E-03 | 8.1052E-03 |
ADE | 3.1455E-03 | 4.6463E-03 | 4.5439E-03 | 4.6656E-03 | 6.0923E-03 | 7.7264E-03 |
CSDE | 9.7295E-03 | 7.3634E-03 | 7.3791E-03 | 7.2028E-03 | 7.1571E-03 | 7.0336E-03 |
ACSDE | 5.2449E-03 | 2.2811E-02 | 2.4206E-02 | 2.4100E-02 | 2.3438E-02 | 2.2736E-02 |
PSO | 1.2374E-02 | 5.3048E-03 | 6.9606E-03 | 6.2216E-03 | 6.6237E-03 | 7.2507E-03 |
FSA | 6.2602E-03 | 3.5832E-03 | 3.5640E-03 | 3.7438E-03 | 4.6444E-03 | 6.0076E-03 |
NN | 3.4667E-02 | 7.3152E-02 | 8.3294E-02 | 9.8081E-02 | 1.1251E-01 | 1.2140E-01 |
LHS2STEPS | 2.1742E-02 | 2.1617E-02 | 2.1475E-02 | 2.1528E-02 | 2.1408E-02 | 2.0572E-02 |
KNNLHS | 4.9637E-03 | 5.3025E-03 | 4.9676E-03 | 5.8222E-03 | 5.9288E-03 | 6.1111E-03 |
Mean L∞ | ||||||
KNN | 4.1213E-02 | 4.3991E-02 | 4.4840E-02 | 4.4840E-02 | 4.9788E-02 | 4,8775E-02 |
LHS | 1.3126E-01 | 1.2567E-01 | 1.2326E-01 | 1.1998E-01 | 1.1904E-01 | 1,2479E-01 |
ALHS | 1.0932E-01 | 1.1062E-01 | 1.1454E-01 | 1.1529E-01 | 1.1662E-01 | 1,2263E-01 |
DE | 2.4302E-02 | 2.7539E-02 | 3.2492E-02 | 3.8239E-02 | 4.4672E-02 | 5,2324E-02 |
ADE | 1.5357E-02 | 2.8079E-02 | 3.2408E-02 | 3.7967E-02 | 4.4269E-02 | 5,1787E-02 |
CSDE | 5.8757E-02 | 6.0849E-02 | 6.3653E-02 | 6.5595E-02 | 6.9252E-02 | 7,2312E-02 |
ACSDE | 3.1167E-02 | 1.4168E-01 | 1.4022E-01 | 1.4177E-01 | 1.4747E-01 | 1,4928E-01 |
PSO | 4.6008E-02 | 4.6052E-02 | 5.2500E-02 | 5.6560E-02 | 6.2328E-02 | 6,7784E-02 |
FSA | 1.6543E-02 | 2.2407E-02 | 2.7837E-02 | 3.3938E-02 | 4.0081E-02 | 4,9525E-02 |
NN | 8.1834E-02 | 2.3886E-01 | 3.0443E-01 | 3.6030E-01 | 3.8488E-01 | 4,1158E-01 |
LHS2STEPS | 1.0144E-01 | 1.0014E-01 | 1.0167E-01 | 1.0810E-01 | 1.1415E-01 | 1,1607E-01 |
KNNLHS | 2.7072E-02 | 3.7594E-02 | 4.0961E-02 | 4.3185E-02 | 4.7107E-02 | 5,0309E-02 |
STD L∞ | ||||||
KNN | 2.7190E-02 | 3.0633E-02 | 2.8849E-02 | 2.8849E-02 | 2.8941E-02 | 2.9448E-02 |
LHS | 8.7321E-02 | 8.2894E-02 | 7.8177E-02 | 7.6610E-02 | 7.7353E-02 | 8.1824E-02 |
ALHS | 8.8403E-02 | 8.9334E-02 | 8.8679E-02 | 8.8367E-02 | 8.9894E-02 | 8.7995E-02 |
DE | 2.2486E-02 | 1.8169E-02 | 1.9161E-02 | 1.9406E-02 | 2.4870E-02 | 3.2087E-02 |
ADE | 1.5143E-02 | 1.8220E-02 | 1.9221E-02 | 2.0268E-02 | 2.5709E-02 | 3.0895E-02 |
CSDE | 4.0726E-02 | 2.6558E-02 | 2.8031E-02 | 2.8521E-02 | 2.6937E-02 | 2.9098E-02 |
ACSDE | 2.1241E-02 | 6.5362E-02 | 6.7719E-02 | 7.0399E-02 | 6.9278E-02 | 6.8662E-02 |
PSO | 5.6853E-02 | 2.4197E-02 | 2.5349E-02 | 2.5845E-02 | 2.8182E-02 | 3.1462E-02 |
FSA | 2.3181E-02 | 1.5810E-02 | 1.5634E-02 | 1.6278E-02 | 2.0246E-02 | 2.6014E-02 |
NN | 9.1032E-02 | 1.6232E-01 | 1.8105E-01 | 1.8648E-01 | 2.0080E-01 | 1.9789E-01 |
LHS2STEPS | 7.1699E-02 | 6.9753E-02 | 7.1235E-02 | 7.3554E-02 | 7.4717E-02 | 7.2363E-02 |
KNNLHS | 2.2045E-02 | 2.6180E-02 | 2.4270E-02 | 2.8580E-02 | 2.9683E-02 | 3.2208E-02 |
In general, integration of the LHS algorithm enhanced the accuracy of the KNN model. This improvement can be attributed to the LHS sampling approach, which mitigates the discontinuity issue associated with KNN predictions by sampling around the initial estimate.
Furthermore, the NN method exhibited a decline in accuracy when confronted with noise-contaminated observational backgrounds. This is attributed to its reliance on gradients, rendering it less effective in discrete cases.
To visualize the performance of the inverse algorithms directly, we randomly selected one specimen from a test dataset. Figures (11 and 12) depict the relative L2 error of the field prediction at the 11th vertical level. Readers are referred to Fig. 10 for the true field values.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F011.jpg)
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F012.jpg)
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F010.jpg)
To fully demonstrate the reconstruction accuracy of our approach, we examined cases in which the reconstruction was relatively poor, particularly around the control rod interfaces. Reconstruction of the physical fields in locations near the control rod interfaces is challenging because the behavior in these regions is not easily captured with high fidelity. Readers can refer to Fig. 10 for the true field value on this axis, and Figs. 13 and 14 depict the prediction errors of various methods on this axis.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F013.jpg)
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F014.jpg)
However, our results also show that even in such difficult cases, our method can achieve good reconstruction in most cases. Although some techniques performed worse around the interfaces, the majority of our proposed methods still facilitated a reconstruction quality comparable to other areas across the domain. Overall, this validation indicates that the proposed approaches hold promise for representing complex reactor physics behaviors with suitable accuracy, even when challenges arise locally because of design complexities, such as control rod insertion points.
Evaluation of State Estimation Phase by Time Cost
In this section, we present an overview of the time costs (we take the mean of the 40 specimens) associated with each proposed algorithm in the state-estimation phase, which can be found in Table 3. The optimal values are highlighted in bold, and the second optimal value is underlined.
noise level | 0 | 0.01 | 0.02 | 0.03 | 0.04 | 0.05 |
---|---|---|---|---|---|---|
KNN | ||||||
LHS | 7.3946E-01 | 7.3326E-01 | 7.2510E-01 | 7.3279E-01 | 7.2927E-01 | 7.2180E-01 |
ALHS | 7.1437E-01 | 7.3674E-01 | 7.1443E-01 | 7.1613E-01 | 7.1730E-01 | 7.1050E-01 |
DE | 3.0745E+00 | 2.9459E+00 | 2.8510E+00 | 2.9223E+00 | 2.8299E+00 | 2.7874E+00 |
ADE | 2.9249E+00 | 2.9339E+00 | 2.8915E+00 | 2.9284E+00 | 2.9743E+00 | 3.0290E+00 |
CSDE | 4.7069E+00 | 4.6679E+00 | 4.6691E+00 | 4.6531E+00 | 4.6615E+00 | 4.6227E+00 |
ACSDE | 4.8054E+00 | 4.6563E+00 | 4.6374E+00 | 4.6886E+00 | 4.6303E+00 | 4.7069E+00 |
PSO | 2.4404E+00 | 2.5169E+00 | 2.4969E+00 | 2.3198E+00 | 2.3551E+00 | 2.3151E+00 |
FSA | 2.7848E+00 | 2.8079E+00 | 2.7324E+00 | 2.7212E+00 | 2.7090E+00 | 2.6977E+00 |
NN | 4.1338E-04 | 4.2277E-04 | 4.0678E-04 | 4.0900E-04 | 4.0602E-04 | 4.0516E-04 |
LHS2STEPS | 7.0137E-01 | 7.0546E-01 | 7.0171E-01 | 7.0508E-01 | 7.0827E-01 | 7.0630E-01 |
KNNLHS | 7.0515E-02 | 6.4439E-02 | 6.3790E-02 | 6.4109E-02 | 6.3136E-02 | 6.3097E-02 |
It is worth mentioning that the NN method demonstrated notable speed. However, metaheuristic algorithms incur relatively high time costs. This is primarily attributed to the repeated computation of the cost function in Eq. 12, which occurs 50 times for each population member over 50 iterations/generation cycles, resulting in a total of 2500 cost function evaluations. Moreover, these algorithms involve additional computations and sorting operations for information exchange and identification of the best solutions, contributing to the discrepancy in time cost.
Furthermore, it is important to note that the methods LHS, ALHS, and LHS2STEPS have approximately the same time costs. In the algorithm LHS, we set the number of samples to 1000. In algorithms ALHS and LHS2STEPS, we set the first-stage sampling number to 750, and the second-stage sampling is repeated five times, with each repetition sampling number uniformly set to 50, adding up to 1000 in total. Model-based algorithms, such as KNN and its hybrid KNNLHS, have relatively low time costs, that is, less than one second, which is acceptable in the industry.
Comprehensive Evaluation of Various Algorithms
In the state-estimation phase, a comparison of the proposed algorithms in terms of time cost and accuracy is presented in Fig. 15. The KNNLHS algorithm demonstrates a commendable level of accuracy, achieving solutions within 0.1 s for both clean and noise-contaminated observation backgrounds. Consequently, KNNLHS can be considered suitable for real-time online inverse modeling in the context of RODT. Moreover, in scenarios in which time constraints are not limiting factors, the FSA, DE, and ADE algorithms provide reasonably accurate solutions.
-202408(1)/1001-8042-35-08-008/alternativeImage/1001-8042-35-08-008-F015.jpg)
Conclusion and Future Works
In this study, we developed and implemented a series of powerful algorithms for the RODT, aiming to address the challenges in parameter identification and state estimation for nuclear reactor monitoring and optimization applications. Novel hybrid optimization algorithms, including ADE, CSDE, ACSDE, and KNNLHS, were designed and implemented to effectively solve the inverse problems posed by the discontinuous surrogate forward modeling approach. Both deterministic and metaheuristic optimization methods have been investigated for their applicability to nuclear engineering systems characterized by high-dimensional, nonlinear relationships. The integration of coarse and fine grid searching strikes an optimal balance between computational efficiency and solution accuracy.
Through extensive experimentation and evaluation, we conducted a thorough comparative analysis of parameter identification accuracy, field reconstruction accuracy, and convergence profile. FSA exhibited the fastest convergence rate, and the proposed ADE method follows closely behind. Moreover, integrating standard DE or CSDE algorithms within the SVC coarse grid search framework enables these techniques to effectively escape local optima in early iterations while maintaining relatively low-cost function values, surpassing their standalone implementations. Additionally, it is observed that pairing the LHS approach within the KNN surrogate model enhances the accuracy by mitigating discontinuity issues through local sampling around the initial guess. Thus, our approach effectively addresses the challenges of discontinuity and non-differentiability in surrogate forward mapping generated by the KNN model. Conversely, the NN method experiences declining accuracy when faced with noise-contaminated measurements, likely owing to overreliance on gradient information, which proves less effective under such uncertainty.
Our algorithms demonstrated high computational efficiency, with solutions obtained within 1 s and as fast as
By specifically tailoring solutions to nuclear inverse and surrogate modeling challenges, this research advances the state-of-the-art RODT methodology. The improved parameter identification and state estimation facilitate the optimization of reactor monitoring, control, and performance through data-driven DT applications. In the future, novel methods can contribute to our RODT system, and with the same methodology, we can support the development and implementation of these transformative technologies in other industries requiring real-time DTs.
Continued work validating diverse systems and progressively integrating multi-physics modeling holds promise for fully realizing the potential of RODTs in transforming nuclear engineering practice with powerful real-time and decision-support capabilities. Realizing this potential requires ongoing progress, collaboration, and open innovation across expert areas.
Digital twin in industry: State-of-the-art
. IEEE Transactions on industrial informatics 15, 2405-2415 (2018).A state-of-the-art survey of digital twin: techniques, engineering product lifecycle management and business innovation perspectives
. Journal of Intelligent Manufacturing 31, 1313-1337 (2020). https://doi.org/10.1007/s10845-019-01512-wReview of digital twin about concepts, technologies, and industrial applications
. Journal of Manufacturing Systems 58, 346-361 (2021). https://doi.org/10.1016/j.jmsy.2020.06.017Digital twin: Origin to future
. Applied System Innovation 4, 36 (2021). https://doi.org/10.3390/asi4020036A conceptual framework for ai-based operational digital twin in chemical process engineering
.Engine fleet-management: The use of digital twins from a mro perspective
. (2017) https://doi.org/10.1115/GT2017-63336Virtual, digital and hybrid twins: A new paradigm in data-based engineering and engineered data
. Archives of Computational Methods in Engineering 27, 105-134 (2018). https://doi.org/10.1007/s11831-018-9301-4Hybrid digital twins: A primer on combining physics-based and data analytics approaches
. IEEE Software 39, 47-52 (2022). https://doi.org/10.1109/MS.2021.3134042Online autonomous calibration of digital twins using machine learning with application to nuclear power plants
. Applied Energy 326,Digital twin in energy industry: Proposed robust digital twin for power plant and other complex capital-intensive large engineering systems
. Energy Reports 8, 3704-3726 (2022). https://doi.org/10.1016/j.egyr.2022.02.305Data-enabled physics-informed machine learning for reduced-order modeling digital twin: Application to nuclear reactor physics
. Nuclear Science and Engineering 196, 1-26 (2022). https://doi.org/10.1080/00295639.2021.2014752Research on fault diagnosis algorithm based on convolutional neural network
. Vol. 1, 2019, pp. 8–12. https://doi.org/10.1109/IHMSC.2019.00010An efficient digital twin based on machine learning svd autoencoder and generalised latent assimilation for nuclear reactor physics
. Annals of Nuclear Energy 179,Parameter identification and state estimation for nuclear reactor operation digital twin
. Annals of Nuclear Energy 180,A survey of model reduction methods for large-scale systems
. (2000). https://doi.org/10.1090/conm/280/04630Optimal reduced space for variational data assimilation
. Journal of Computational Physics 379,. https://doi.org/10.1016/j.jcp.2018.10.042Sensor placement in nuclear reactors based on the generalized empirical interpolation method
. J. Comput. Phys. 363, 354-370 (2018). https://doi.org/10.1016/j.jcp.2018.02.050Model reduction and approximation: theory and algorithms
, (A survey of projection-based model reduction methods for parametric dynamical systems
. SIAM Review 57, 483-531 (2015). https://doi.org/10.1137/130932715Toward predictive digital twins via component-based reduced-order models and interpretable machine learning
. 2020. https://doi.org/10.2514/6.2020-0418Pbdw: A non-intrusive reduced basis data assimilation method and its application to an urban dispersion modeling framework
. Applied Mathematical Modelling 76, 1-25 (2019). https://doi.org/10.1016/j.apm.2019.05.012Digital twin: Values, challenges and enablers from a modeling perspective
. IEEE Access 8, 21980-22012 (2020). https://doi.org/10.1109/ACCESS.2020.2970143Real time parameter identification and solution reconstruction from experimental data using the proper generalized decomposition
. Computer Methods in Applied Mechanics and Engineering 296, 113-128 (2015). https://doi.org/10.1016/j.cma.2015.07.020Reduced order modeling and parameter identification of a building energy system model through an optimization routine
. Applied Energy 162, 1010-1023 (2016). https://doi.org/10.1016/j.apenergy.2015.10.137Pod/deim reduced-order modeling of time-fractional partial differential equations with applications in parameter identification
. Journal of Scientific Computing 74, 220-243 (2016). https://doi.org/10.1007/s10915-017-0433-8Parameter identification of reduced-order electrochemical model simplified by spectral methods and state estimation based on square-root cubature kalman filter
. Journal of Energy Storage 46,A generalized empirical interpolation method: Application of reduced basis techniques to data assimilation
. Springer INdAM Series 4,.A parameterized-background data-weak approach to variational data assimilation: formulation, analysis, and application to acoustics
. International Journal for Numerical Methods in Engineering 102, 933-965. https://doi.org/10.1002/nme.4747Data assimilation in reduced modeling. SIAM/ASA J. Uncertain
. Quantification 5, 1-29 (2015). https://doi.org/10.1137/15M102538Data assimilation with reduced basis and noisy measurement: Applications to nuclear reactor cores
. Ph.D. thesis,Optimal reduced space for variational data assimilation
. Journal of Computational Physics 379, 51-69 (2019). https://doi.org/10.1016/j.jcp.2018.10.042Optimal and fast field reconstruction with reduced basis and limited observations: Application to reactor core online monitoring
. Nuclear Engineering and Design 377,Sensitivity and uncertainty analysis of a reduced-order model of nonlinear bwr dynamics: I. forward sensitivity analysis
. Annals of Nuclear Energy 148,Multiobjective optimal control methods for the navier-stokes equations using reduced order modeling
. Acta Applicandae Mathematicae 161, 171-199 (2018). https://doi.org/10.1007/s10440-018-0209-7Reduced order methods for uncertainty quantification problems
. ETH Zurich, SAM Report 3,.Validation and uncertainty quantification of multiphase-cfd solvers: A data-driven bayesian framework supported by high-resolution experiments
. Nuclear Engineering and Design 354,A weighted pod-reduction approach for parametrized pde-constrained optimal control problems with random inputs and applications to environmental sciences
. Comput. Math. Appl. 102, 261-276 (2021). https://doi.org/10.1016/j.camwa.2021.10.020An efficient computational framework for naval shape design and optimization problems by means of data-driven reduced order modeling techniques
. Bollettino dell’Unione Matematica Italiana 14, 211-230 (2020). https://doi.org/10.1007/s40574-020-00263-4Koopman-based approach to nonintrusive reduced order modeling: Application to aerodynamic shape optimization and uncertainty propagation
. AIAA Journal 58, 2221-2235 (2020). https://doi.org/10.2514/1.j058744Reduced order modeling for nonlinear pde-constrained optimization using neural networks
.Reduced order modeling for time-dependent optimization problems with initial value controls
. SIAM J. Sci. Comput. 40,. https://doi.org/10.1137/16M1109084A simulation app based on reduced order modeling for manufacturing optimization of composite outlet guide vanes
. Advanced Modeling and Simulation in Engineering Sciences 4, 1-26 (2017). https://doi.org/10.1186/s40323-017-0087-yMultiobjective pde-constrained optimization using the reduced-basis method
. Advances in Computational Mathematics 43, 945-972 (2017). https://doi.org/10.1007/s10444-016-9512-xMl-knn: A lazy learning approach to multi-label learning
. Pattern Recognit. 40, 2038-2048 (2007). https://doi.org/10.1016/j.patcog.2006.12.019Top 10 algorithms in data mining
. Knowledge and Information Systems 14, 1-37 (2007). https://doi.org/10.1007/s10115-007-0114-2Learning k for knn classification
. ACM Trans. Intell. Syst. Technol. 8,. https://doi.org/10.1145/2990508Surrogate and reduced-order modeling: a comparison of approaches for large-scale statistical inverse problems
. Large-Scale Inverse Problems and Quantification of Uncertainty (2010). https://doi.org/10.1002/9780470685853.ch7Development and validation of reactor nuclear design code corca-3d
. Nuclear Engineering and Technology 51, 1721-1728 (2019). https://doi.org/10.1016/j.net.2019.05.015Study of rod ejection accident at hot zero power condition in a pwr using relap5
. Progress in Nuclear Energy 144,Reduced basis method for the rapid and reliable solution of partial differential equations
. Proceedings oh the International Congress of Mathematicians 3, 1255-1270 (2006). https://doi.org/10.4171/022-3/60Efficient reduced-basis treatment of nonaffine and nonlinear partial differential equations
. ESAIM: Mathematical Modelling and Numerical Analysis 41, 575-605 (2007). https://doi.org/10.1051/m2an:2007031The approximation of one matrix by another of lower rank
. Psychometrika 1, 211-218 (1936). https://doi.org/10.1007/BF02288367Solving inverse problems using data-driven models
. Acta Numerica 28, 1-174 (2019). https://doi.org/10.1007/BF02288367Inverse optimization
. Operations Research 49,. https://doi.org/10.1287/opre.49.5.771.10607Direct search methods: then and now
. Journal of Computational and Applied Mathematics 124, 191-207 (2000). https://doi.org/10.1006/jpdc.1997.1409Latin hypercube sampling and the propagation of uncertainty in analyses of complex systems
. Reliability Engineering & System Safety 81, 23-69 (2003). https://doi.org/10.1016/S0951-8320(03)00058-9Bound constraints handling in differential evolution: An experimental study
. Swarm and Evolutionary Computation 50,Constraint handling techniques for metaheuristics: a state-of-the-art review and new variants
. Optimization and Engineering 2251–2298 (2023). https://doi.org/10.1007/s11081-022-09782-9The automatic dynamic penalisation method (adp) for handling constraints with genetic algorithms
. Computer Methods in Applied Mechanics and Engineering 256, 70-87 (2013). https://doi.org/10.1016/j.cma.2012.12.009Differential evolution: A survey of the state-of-the-art
. IEEE Transactions on Evolutionary Computation 15, 4-31 (2011). https://doi.org/10.1109/TEVC.2010.2059031Differential evolution - a simple and efficient heuristic for global optimization over continuous spaces
. Journal of Global Optimization 11, 341-359 (1997). https://doi.org/10.1023/A:1008202821328Proceedings of the IEEE international conference on neural networks, Particle swarm optimization
. Vol. 4,Stochastic relaxation, gibbs distributions, and the bayesian restoration of images
. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6, 721-741 (1984). https://doi.org/10.1109/TPAMI.1984.47675962021 IEEE Statistical Signal Processing Workshop (SSP)
, Simulated annealing: a review and a new scheme. 2021, pp. 101-105. https://doi.org/10.1109/SSP49050.2021.9513782Evolutionary simulated annealing for fuel loading optimization of vver-1000 reactor
. Annals of Nuclear Energy 151,Fast simulated annealing
. Phys. Lett. A 122, 721-741 (1984). https://doi.org/10.1016/0375-9601(87)90796-1World congress on nature & biologically inspired computing (NaBIC), Cuckoo search via lévy flights
.Artificial neural networks: A tutorial
. Computer 29, 31-44 (1996). https://doi.org/10.1109/2.485891Combining data assimilation and machine learning to emulate a dynamical model from sparse and noisy observations: A case study with the lorenz 96 model
. Journal of Computational Science 44,A data-enabled physics-informed neural network with comprehensive numerical study on solving neutron diffusion eigenvalue problems
. Annals of Nuclear Energy 183,Support vector machines: Optimization and validation for land cover mapping using aerial images and lidar data
. (2011). https://www.isprs.org/proceedings/2011/isrse-34/211104015Towards automatic freeform optics design: coarse and fine search of the three-mirror solution space
. Light, Science Applications 10, 65 (2021). https://doi.org/10.1038/s41377-021-00510-zRandom search for hyper-parameter optimization
. Journal of machine learning research 13,.Mutations as levy flights
. Scientific Reports 11,. https://doi.org/10.1038/s41598-021-88012-1The impact of class imbalance in classification performance metrics based on the binary confusion matrix
. Pattern Recognition 91, 216-231 (2019). https://doi.org/10.1016/j.patcog.2019.02.023177 core nuclear design for hpr1000
. Nucl. Power Eng. 40, 0008-00012 (2019). https://doi.org/10.13832/j.jnpe.2019.S1.0008 (in Chinese)Development and validation of reactor nuclear design code corca-3d
. Nuclear Engineering and Technology 51, 1721-1728 (2019). https://doi.org/10.1016/j.net.2019.05.015The authors declare no conflict of interest.