Introduction
In the nuclear engineering domain, the fundamental mode solution of the K-eigenvalue problem based on the steady-state multigroup neutron diffusion theory is crucial for simulation and analysis of nuclear reactors. The eigenvalue equation can be expressed as follows:
As mentioned above, many traditional numerical methods were proposed to solve neutron diffusion equations. However, a dense mesh is required to ensure highly accurate results. The model consumes more computational resources if the mesh is extremely dense. Inaccurate solutions are obtained if the mesh is coarse. Moreover, for high-dimensional problems, classical methods are either less efficient or not successful due to the problems involving dimensionality. A neural network can potentially enhance its performance by using a capable hypothesis space due to its relatively low statistical error [13].
Neural networks exhibit multiple potential benefits in solving nuclear engineering problems when compared with conventional numerical methods.
• They provide mesh-free solutions to approximate physics fields in nuclear reactors, in which mesh generation is significantly complex due to high heterogeneity of geometry and material.
• They provide a general framework to solve high-dimensional problems governed by parameterized PDEs, particularly for original neutron transport equation.This equation comprises seven variables: three spatial variables, two directional variables, one energy variable, and one temporal variable.
• They are able to seamlessly incorporate prior data (potentially with noise) into existing algorithms given that data assimilation is necessary as a post-process for conventional numerical methods [14-19].
• They provide a general framework to solve inverse problems with hidden physics. Conversely, they are typically prohibitively expensive and require different formulations and elaborate computer codes for conventional numerical methods.
In recent years, neural networks have been widely used to solve Partial Differential Equations (PDEs) [20] and achieved remarkable success.
Based on neural networks, a large number of methods were proposed to solve PDE, such as the deep backward stochastic differential equation (BSDE) method [21, 22], deep Galerkin method (DGM) [23], deep Ritz method (DRM) [24], and Physics-Informed Neural Network (PINN) [25]. The deep BSDE method reformulates PDEs using backward stochastic differential equations and approximates the gradient of an unknown solution using neural networks. Although DGM and PINN appear independently under two names in the literature, they are similar. They both train neural networks by minimizing the mean squared error loss of the residual equation. However, DRM reformulates the original problem into a variational problem and trains neural networks by minimizing the energy function of the variational problem. Specifically, DRM and PINN [25] attracted widespread attention. Extensive studies focused on these methods to solve a variety of problems, including fiber optics [26], hyperelasticity [27], solid mechanics [28], heat transfer problems [29-31], inverse problems [32-35], and uncertainty quantification [36-39]. However, a few studies focused on solving eigenvalue problems [24, 40-45].
The need to solve eigenvalue problems can be traced back to 2018 [24]. A deep Ritz method to solve variational problems was proposed, and several examples elucidated on how to use DRM to solve eigenvalue problems [24]. First, the original eigenvalue problem was transformed into a variational problem. Then, a specially defined loss function was constructed, termed as the Rayleigh quotient, using the variational principle [46]. The Rayleigh quotient is a well-known approximation of the eigenvalue of matrix 𝑨, which is defined by
In a recent study, PINN was applied to solve neutron diffusion equations [51, 45], where the authors used a free learnable parameter to approximate the eigenvalue and a novel regularization technique to exclude null solutions from the PINN framework. A conservative physics-informed neural network (cPINN) was proposed in discrete domains for nonlinear conservation laws [52]. Moreover, cPINN [53] was applied to solve heterogeneous neutron diffusion problems in one-dimensional cases, which develops PINN for each subdomain and considers additional conservation laws along the interfaces of subdomains (a general consideration in reactor physics [11]), which is involved in neural network training as the variable to be optimized.
More recently, a data-enabled physics-informed neural network (DEPINN) [40] was proposed to solve neutron diffusion eigenvalue problems. To achieve acceptable engineering accuracy for complex engineering problems, it is suggested that a very small amount of prior data from physical experiments be used to improve the training accuracy and efficiency. In contrast to PINN, which solves the neutron diffusion eigenvalue problem directly, an autoencoder-based machine learning method in combination with the reduced-order method [54, 55] was proposed [56] to solve the same problem. However, it still relies on solving governing equations with traditional numerical methods such as the finite difference method.
Although DRM provides a way to solve eigenvalue problems with neural networks, as shown in Sect. 4.2.2, results indicate that DRM is not stable when solving two-dimensional cases. First, DRM learns the eigenvalue and eigenfunction at the early stage of the training process. Subsequently, DRM attempts to learn a smaller eigenvalue after it is close to the true eigenvalue. Finally, DRM successfully learns one smaller eigenvalue that may be close to the true eigenvalue and one incorrect eigenfunction that is meaningless and far from the true eigenfunction. Additionally, the framework of a neural network based on the power method [50] is unsuitable for generalized eigenvalue problems. Therefore, it is necessary to propose a new algorithm to solve K-eigenvalue problems.
The study focuses on eigenvalue problems, which are also interface problems in which the eigenfunctions are continuous on the interface, and the derivatives of the eigenfunction are not continuous at the interface. Specifically, in the nuclear reactor physics domain, this is a general problem in which the reactor core is composed of fuel assemblies of different fissile nuclides enrichments [1]. Some studies focused on the use of neural networks to solve elliptic interface problems. Some researchers [57] use the idea of DRM and formulated the PDEs into the variational problems, which can be solved using the deep learning approach. They presented a novel mesh-free numerical method for solving elliptical interface problems based on deep learning [58]. They employed different neural networks in different subdomains and reformulated the problem as a least-squares problem. A similar case exists, in which the authors [53] enforce the interface conditions using piecewise neural network. In contrast to the methods, a discontinuity-capturing shallow neural network (DCSNN) [59] has been proposed for elliptic interface problems. The crucial concept of DCSNN is that a d-dimensional piecewise continuous function can be extended to a continuous function defined in (d+1)-dimensional space, where the augmented coordinate variable labels the pieces of each subdomain. However, to the best of the authors’ knowledge, only a few studies focused on the use of neural networks to solve eigenvalue problems that incorporate interface problems involving regions of different materials. However, challenges exist at least on three fronts.
• Designing a neural network that is more suitable for K-eigenvalue problems for more complicated/medium-size test problems.
• Dealing with the interface problem in a more general and understandable manner when designing the neural network.
• Proposing a framework effectively enhances the robustness of the neural network and improves the efficiency of utilizing the noisy prior data.
To address the aforementioned challenges and advance beyond the state-of-the-art research [51, 45, 53], we initially introduced the study [40]. This served as a preliminarily demonstration of the applicability of the PINN approach to reactor physics eigenvalue problems in complex engineering scenarios. The contributions of this study are as follows.
• Firstly, we extend the Inverse Power Method Neural Network (IPMNN)[49] to the so-called Generalized Inverse Power Method Neural Network (GIPMNN) to solve the smallest eigenvalue and the related eigenfunction of the generalized K-eigenvalue problems. Compared to DEPINN from our previous study [53], we omit the prior data in the training process and attempt to solve the K-eigenvalue problems from a data-free mathematical/numerical perspective.
• Then, we propose a Physics Constrained GIPMNN (PC-GIPMNN) to address the interface problem in a more general and understandable manner with respect to previous studies [51, 45, 53].
• Finally, we conduct a thorough comparative study of GIPMNN, PC-GIPMNN, and DRM using a variety of numerical tests. We evaluate the applicability and accuracy of these three methods using typical 1D and 2D test cases in reactor physics, particularly accounting for material discontinuities in different geometries. In the 1D example, we determine the optimal ratio of outer to inner iterations, a finding that may be particularly relevant for GIPMNN. Additionally, we observe the failure of DRM in the 2D experiments, whereas PC-GIPMNN consistently outperforms GIPMNN and DRM over a set number of epochs.
The rest of this paper is organized as follows. The governing equations for the eigenvalue problems are presented in Sect. 2. In Sect. 3, we propose GIPMNN and PC-GIPMNN and introduce DRM in our cases. In Sect. 4, the results of 1D and 2D test cases are listed to verify the three methods. Finally, conclusions and future research are discussed in Sect. 5.
K-eigenvalue Problems
This section introduces the equations that govern the neutron criticality over a spatial domain. We recall Eqs. (1) and (2). The generalized K-eigenvalue problem can be formulated as follows:
Two main boundary conditions are imposed on the diffusion equation. One condition represents a surface on which neutrons are reflected back into the domain (reflective condition), and the other condition represents surfaces that allow neutrons to escape from the system (vacuum or bare condition). Both the conditions are satisfied by relating the flux solution to its gradient on the boundary, i.e.,
PINN as a eigenvalue solver
In this subsection, we discuss using PINN to solve the generalized eigenvalue problem (2).
The eigenfunction of the operator
As mentioned previously, we are concerned with the smallest eigenvalue and associated eigenfunction. Hence, the eigenvalue (Rayleigh quotient) is viewed as a loss term in PINN to attain the lowest eigenvalue.
Finally, the total loss function (10) in PINN corresponds to the weighted sum of the two objectives (8) and (9), where α and β corresponds to the weights.
Unfortunately, PINN does not work for the cases in the study and even performs worse than DRM. Therefore, we only compared the results of our method with those of DRM.
Methodologies
In this section, we extend our previous study [49] and discuss the use of a neural network to numerically solve the smallest eigenvalue problem. The main concept is to use a neural network to approximate the eigenfunction and compute the eigenvalue based on the Rayleigh quotient using an eigenvector expressed by the points calculated using the eigenfunction.
Neural Network Architecture
Next, we discuss the neural network structure used to approximate eigenfunction ϕ(𝒙). The neural network architecture employed in the study is the same as ResNet [60], which is built by stacking several residual modules. It is one of the most popular models used in Deep Learning, and it is also commonly used in the field for solving PDEs via neural networks [24, 61, 62]. Each module has one skip connection, and each block consists of two fully connected layers as shown in Figure 1.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F001.jpg)
In the network, let
Subsequently, the neural network
Recap of Inverse Power Method Neural Network
We consider the following linear eigenvalue problem:
Inspired by the idea of the inverse power method, IPMNN [49] was proposed to solve for the smallest eigenvalue and associated eigenfunction. Equation (16) depicts the key step of the inverse power method, and equation (17) is analogous to equation (16), where
Here, the neural network
Generalized Inverse Power Method Neural Network
In standard nuclear engineering procedures, we normally employ the inverse power method to solve for the smallest eigenvalue and associated eigenvector, which relies on the discretization of Equation (3). IPMNN [49] was proposed to solve the smallest eigenvalue problem, which is a mesh-free method realized by a neural network. However, the method is restricted to solving the equation
Given that N is the number of points for training the neural network, Nb denotes the number of points on the boundary Ω, Nepoch denotes the maximum number of epochs, λ0 denotes the initial guess of the eigenvalue, and the stopping criterion ϵ. | ||
Step 1: Build data set S for loss function Lossgipmnn and data set Sb for loss function Lossb. | ||
Step 2: Initialize a neural network with random initialization of parameters. | ||
for |
||
Input all points in S and Sb into neural network |
||
Let |
||
Update parameters θ of the neural network via gradient descent. | ||
if Lossdrm<∈ then | ||
Record the best eigenvalue and eigenfunction. | ||
If the stopping criterion is met, the iteration can be stopped. | ||
end | ||
end |
First, the generalized inverse power method is used to solve Eq. (20). The key step to focus on is given in Eq. (21), where
We use a neural network
In GIPMNN, Eq. (22) is an analog of Eq. (21), where
The Neumann and Robin boundary conditions were used for the eigenvalue problem. It is difficult to enforce them by encoding the boundary conditions into a neural network, as in [49, 63, 64], where the Dirichlet and periodic boundary conditions are used.
We form the loss function Lossb in Eq. (8), where Nb denotes the number of sampling points on Ω, and 𝒙i is a point in the sampling set
Therefore, the loss function is defined in Equations (24) and (25) as follows: where the surface indicates that neutrons are reflected back into the domain or that the surface allows neutrons to escape from the system.
The total loss function (26) denotes the weighted sum of the objectives (23) and (8). For the process of GIPMNN, refer to Algorithm 1.
Physics Constrained Generalized Inverse Power Method Neural Network
Although GIPMNN can solve the eigenvalue problems of Eq. (3), it is still difficult to solve eigenvalue problems with discontinuous coefficients in different regions. We discuss interface problems, which implies that the eigenfunction may be continuous at the interface, and the derivatives of the eigenfunction may not be continuous at the interface. The enforcement of interface conditions is very important for GIPMNN.
In the study, inspired by the idea of a piecewise deep neural network [58], we propose a PC-GIPMNN to solve eigenvalue problems with interface conditions. However, instead of employing different neural networks in different subdomains, we use only one neural network and multiple neurons in the output layer, as shown in Fig. 2. It should be noted that each neuron in the output layer corresponds to a subdomain. Therefore, we can obtain outputs in different subdomains that can be used to enforce the conditions at the interface.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F002.jpg)
Suppose that there are two domains Ωl and Ωr, with an interface Γ, which is the cross region between the two domains. Given the properties of the neutron population ϕ(𝒙), we can summarize that the eigenfunction will satisfy two interface conditions, i.e., (27) and Eq. (28), where ϕl and ϕr represent the eigenfunctions defined in Ωl and Ωrm, respectively, and 𝒏 denotes the normal vector pointing from Ωr to Ωl. Dl and Dr are the coefficients defined in Ωl and Ωr, respectively, which are discontinuous at the interface. Equation (27) indicates that the eigenfunction is continuous at the interface, i.e., the neutron flux is continuous at the interface. Equation (28) indicates that neutron flow is continuous at the interface.
Assume that SΓ corresponds to the set of points at the interface Γ and
By combining (26), (29), and (30), Equation (31) is the total loss defined in our method PC-GIPMNN, where α, β, γ, and δ are the weights of the different losses. In subsequent experiments, we chose 1. In the study, we focused on the proposed algorithms and neglected the influence of weights. We are expecting that our proposed algorithms are universal and achieve better results, even without adjusting the weights. Therefore, we select all weights as 1.
Moreover, the eigenfunction can be represented as:
Deep Ritz Method
DRM is a deep-learning-based method for numerically solving variational problems [24]. It reformulates the original PDEs into equivalent variational equations and defines the loss function based on variational formulations. The solutions of PDEs are represented by a neural network, and the derivatives are calculated using AD. DRM [24] is also used to solve eigenvalue problems. Furthermore, we specify how to use DRM to solve Eq. (3).
We consider the variational principle of the smallest eigenvalues.
After the optimal approximation is obtained by solving the optimization problem (31), we obtain the smallest eigenvalue
Given N denotes the number of points for training the neural network, Nb denotes the number of points on the boundary Ω, Length denotes a measure of Ω, Nepoch denotes the maximum number of epochs, and ϵ denotes the stopping criterion | ||
Step 1: Build data set S for the loss of the Rayleigh quotient and data set Sb for the loss of boundary. | ||
Step 2: Initialize a neural network with random initialization of parameters. | ||
for |
||
Input all points in S and Sb into neural network |
||
Let |
||
Update parameters θ of the neural network via gradient descent. | ||
if Lossdrm < ϵ then | ||
Record the best eigenvalue and eigenfunction. | ||
If the stopping criterion is met, then the iteration can be stopped. | ||
end | ||
end |
Experiments
In this section, we present numerical experiments to compare the applicability and accuracy of GIPMNN, PC-GIPMNN, and DRM for solving the smallest eigenvalue problems in reactor physics. In all the experiments below, we chose the Adam optimizer with an initial learning rate 10-3 to minimize the loss function. Furthermore, we trained the neural network with the architecture of ResNet on a server equipped with CentOS 7 system, one Intel Xeon Platinum 8358 2.60-GHz CPU, and one NVIDIA A100 80GB GPU. Moreover, unless otherwise specified, the activation function was selected as the tanh function.
One-Dimensional Slab Reactor
We consider one-dimensional case with a simple slab reactor consisting of a domain bounded by vacuum or bare surfaces, as shown in Fig. 3. The length of each slab reactor was 10 cm. It consists of fuel and control rod regions labeled 1, 2, and 3. The control rods were located between 2.2 and 2.5 cm and between 7.5 and 7.8 cm on the x-axis.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F003.jpg)
As shown in Fig. 3, there were two control rods in the one-dimensional slab reactor. Either device can be withdrawn or inserted. Three scenarios were designed to model the reactor and completely simulate the actions of the control rods. The three situations are labeled F1, F2, and F3 in Table 1. They indicate that both rods are withdrawn, only the left rod is inserted, and only the right rod is inserted. We also considered three other problems, R1, R2, and R3 in Table 1. Problem R1 is the same as F1, which is designed to simulate the withdrawal of both control rods. Here, we use R1 to facilitate comparison with R2 and R3. Although problems R2 and R3 denote that all the control rods are inserted, problem R2 resembles heavily inserted control rods, and problem R3 resembles slightly inserted control rods.
Problem | Region | ∑a(cm-1) | ∑s(cm-1) | v ∑f(cm-1) |
---|---|---|---|---|
F1 | Region1 | 0.4 | 2.0 | 0.5 |
Region2 | 0.4 | 2.0 | 0.5 | |
Region3 | 0.4 | 2.0 | 0.5 | |
F2 | Region1 | 0.4 | 2.0 | 0.5 |
Region2 | 0.6 | 2.0 | 0.3 | |
Region3 | 0.4 | 2.0 | 0.5 | |
F3 | Region1 | 0.4 | 2.0 | 0.5 |
Region2 | 0.4 | 2.0 | 0.5 | |
Region3 | 0.6 | 2.0 | 0.3 | |
R1 | Region1 | 0.4 | 2.0 | 0.5 |
Region2 | 0.4 | 2.0 | 0.5 | |
Region3 | 0.4 | 2.0 | 0.5 | |
R2 | Region1 | 0.4 | 2.0 | 0.5 |
Region2 | 0.6 | 2.0 | 0.3 | |
Region3 | 0.6 | 2.0 | 0.3 | |
R3 | Region1 | 0.4 | 2.0 | 0.7 |
Region2 | 0.5 | 2.0 | 0.4 | |
Region3 | 0.5 | 2.0 | 0.4 |
To generate the necessary data to validate the accuracy of GIPMNN, PC-GIPMNN, and DRM, FreeFEM [67-73] is utilized to solve these problems in Table 1. FreeFEM is a partial differential equation solver for non-linear multi-physics systems using the finite element method. We choose number of cells in the x-direction N = 1001 and mesh size Δ x=10-3. The solution of the baseline is obtained by FEM, and uniform cells are used to train GIPMNN, PC-GIPMNN, and DRM.
Using GIPMNN to Solve for Eigenvalue
In this one-dimensional example, we select 20 neurons for each hidden layer in ResNet and Nepoch=50000. Without a loss of generality, the weights α and β of the different losses are not adjusted to achieve better results. Therefore, α and β were set to one.
In Eq. (21), we must solve for
for |
|||
for |
|||
Input all points in S and Sb into neural network |
|||
Let |
|||
if |
|||
end | |||
Update parameters θ of the neural network via gradient descent. | |||
end | |||
if Losstotal < ϵ then | |||
Record the best eigenvalue and eigenfunction. | |||
If the stopping criterion is met, then the iteration can be stopped. | |||
end | |||
end |
We chose the ratios of the outer and inner iterations as 1:1, 1:10, 1:100, 1:1000, and 1:10000 to investigate the effect of the ratio on the results. Ratio 1:n implies that the outer code is executed once, whereas the inner code is executed n times. For comparison, the total number of iterations was fixed at Ntotal=100000 and
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F004.jpg)
Moreover, we fixed the number of outer iterations Nouter=1000 and retrained the neural network using the ratios. The results are shown in Fig. 5. The opposite results are observed when compared with the results in Fig. 4. The ratio 1:1 is the worst ratio because increases in inner iterations lead to a better approximation of the eigenfunction in the next outer iteration. This is consistent with the results of the inverse power method.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F005.jpg)
Using PC-GIPMNN to Solve for Eigenvalue
We divided the slab reactor into five parts from left to right. The output layer included five neurons, and five functions were defined in different subdomains. Here, u, w, and q denote the functions defined in Region 1, v and p are those defined in Regions 2 and 3, respectively.
As shown in Figure 3, the four points are denoted as xi1=2.2, xi2=2.5, xi3=7.5, and xi4=7.8. The interface conditions (29) and (30) are implemented as loss functions defined in (38) and (39) in the 1D example.
Using DRM to Solve for Eigenvalue
The configuration of DRM is identical to that of GIPMNN. Given that it is a variational formulation, and the boundary condition is incorporated into the Rayleigh quotient, we do not use the penalty term to enforce the boundary condition. The loss function Lossdrm is defined as:
Results
The results for the one-dimensional example are listed in Tables 1, 2, 3, and 6. The values of keff and their relative errors are listed in Table 2. For problems F1 and R1, which have continuous coefficients, the results obtained by GIPMNN are better than those obtained by PC-GIPMNN and DRM. For problem F2, the relative error of keff obtained by DRM is better than those obtained by other methods. For other problems with discontinuous coefficients, the results obtained using the PC-GIPMNN are better than those obtained by GIPMNN and DRM. The relative error of keff computed by PC-GIPMNN is approximately 10-5, which is lower than 10-4 as computed by DRM. The relative errors in the eigenfunctions are listed in Table 3. For all problems, PC-GIPMNN can attain better results than GIPMNN and DRM.
Problem | keff (FEM) | keff (GIPMNN) | keff (PC-GIPMNN) | keff (DRM) | |||
---|---|---|---|---|---|---|---|
F1 | 1.2127 | 1.2127 | 1.2127 | 1.2128 | 2.5142×10-06 | 3.7743×10-06 | 5.8568×10-05 |
F2 | 1.1973 | 1.1970 | 1.1973 | 1.1972 | 2.3429×10-04 | 2.6246×10-04 | 1.1679×10-04 |
F3 | 1.1973 | 1.1972 | 1.1973 | 1.1971 | 5.5503×10-05 | 1.6898×10-05 | 1.6809×10-04 |
R1 | 1.2127 | 1.2127 | 1.2127 | 1.2128 | 2.5142×10-06 | 3.7743×10-06 | 5.8568×10-05 |
R2 | 1.1715 | 1.1697 | 1.1715 | 1.1707 | 1.5629×10-03 | 3.2020×10-05 | 7.2244×10-04 |
R3 | 1.6498 | 1.6484 | 1.6498 | 1.6487 | 8.5897×10-04 | 2.9838×10-05 | 6.7968×10-04 |
Problem | ϕrel (GIPMNN) | ϕrel (PC-GIPMNN) | ϕrel (DRM) |
---|---|---|---|
F1 | 2.7301×10-03 | 6.3620×10-04 | 2.6313×10-03 |
F2 | 2.6279×10-02 | 1.6140×10-03 | 8.2217×10-03 |
F3 | 2.7403×10-02 | 1.0120×10-03 | 6.2428×10-03 |
R1 | 2.7301×10-03 | 6.3620×10-04 | 2.6313×10-03 |
R2 | 2.5768×10-02 | 7.5946×10-04 | 6.8807×10-03 |
R3 | 1.4559×10-02 | 7.5257×10-04 | 8.1038×10-03 |
Problem | keff (FEM) | keff (DRM) | ϕrel (DRM) | |
---|---|---|---|---|
F1 | 1.0118 | 1.0423 | 3.0156E-02 | 16.6580 |
F2 | 1.0052 | 1.0399 | 3.4538E-02 | 21.8882 |
F3 | 1.0000 | 1.0428 | 4.2782E-02 | 29.5368 |
F4 | 1.0079 | 1.0533 | 4.5361E-02 | 26.4549 |
F5 | 1.0052 | 1.0450 | 3.9598E-02 | 32.0516 |
R1 | 1.0118 | 1.0423 | 3.0156E-02 | 16.6580 |
R2 | 1.0052 | 1.0399 | 3.4538E-02 | 21.8882 |
R3 | 0.9921 | 1.0399 | 4.8250E-02 | 166.2996 |
R4 | 1.0017 | 1.0397 | 3.7884E-02 | 163.6593 |
R5 | 0.9780 | 1.0498 | 7.3430E-02 | 182.7224 |
R6 | 0.9668 | 1.0492 | 8.5210E-02 | 87.2200 |
R7 | 1.1018 | 1.1517 | 4.5337E-02 | 36.6777 |
In Fig. 6, the eigenfunctions of GIPMNN, PC-GIPMNN, and DRM are compared with those of FEM. Specifically, we follow the conventional normalization process in nuclear reactor physics [1], where the normalization constant is generally computed to make the average reactor power equal to unity; thus, the eigenfunctions are normalized by Eq. (38), where N denotes the number of training points in the entire domain.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F006.jpg)
As reported in a previous study, [49], IPMNN can attain more accurate eigenvalues when compared to those obtained with DRM when linear eigenvalue problems without interface are considered. Therefore, IPMNN and GIPMNN are suitable to determine the eigenfunction of a problem with a strong form, and DRM is similar to FEM in that DRM is applicable for finding the eigenfunction of a problem with a weak form. Consequently, DRM is better than GIPMNN in this one-dimensional example with discontinuous coefficients. However, given that the interface conditions are well implemented in PC-GIPMNN, it successfully learns the eigenvalue and eigenfunction and achieves better results than GIPMNN and DRM, as shown in Tables 2 and 3.
Two-Dimensional Reactor
As shown in Figure 7, a two-dimensional reactor is modeled in a square-shaped domain with 90-cm sides. The reactor was surrounded by a neutron reflector with graphite material, which implies that Robin boundary condition was selected. The main bulk of the reactor corresponds to the fuel region. Within its central region, four control rods can be inserted or withdrawn. When the control rods are withdrawn, materials in the regions are replaced with water, which corresponds to common practice in many reactor designs. Table 4 lists five different materials used in the mock reactor. However, the two types of fuels in Table 4 are designed to simulate different fuel materials. Fuel type 1 defines the standard fuel used in most problems, and fuel type 2 defines an adjusted fuel composition that is different from fuel type 1. Fuel type 2 in problem R7 was used to test whether our methods can be affected by different types of fuels.
Material | ∑a(cm-1) | ∑s(cm-1) | v ∑f(cm-1) |
---|---|---|---|
Fuel 1 | 0.075 | 0.53 | 0.079 |
Fuel 2 | 0.072 | 0.53 | 0.085 |
Water | 0.01 | 0.89 | 0.0 |
Control rod | 0.38 | 0.2 | 0.0 |
Graphite | 0.15 | 0.5 | 0.0 |
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F007.jpg)
As shown in Table 5, there are 12 problems in validating the accuracy of the proposed method. Five full models which are labeled as F1–F5, were proposed to simulate the reactor with all control rods removed and then with only one control rod inserted in the regions of the control rods. As previously discussed, when the control rod was removed, the material in the region was replaced with water. Therefore, W is used to denote water in Table 5. Other seven reactor configurations, denoted by R1-R7, were proposed to simulate the cases where more control rods were inserted. Problems R1 and R2 are equivalent to the full model problems F1 and F2: Problems R3–R6 utilized different combinations of inserted and rejected control rods. It should be noted that in problem R7, the material configuration differs from the material configuration of other problems. The fuel type was replaced with fuel type 2, and the control rods were assumed to be partially inserted, which implies that the materials in the regions corresponded to a mix of control rods and water materials.
Problem | Fuel type | C-R1 | C-R2 | C-R3 | C-R4 |
---|---|---|---|---|---|
F1 | Fuel 1 | W | W | W | W |
F2 | Fuel 1 | CR | W | W | W |
F3 | Fuel 1 | W | CR | W | W |
F4 | Fuel 1 | W | W | CR | W |
F5 | Fuel 1 | W | W | W | CR |
R1 | Fuel 1 | W | W | W | W |
R2 | Fuel 1 | CR | W | W | W |
R3 | Fuel 1 | W | CR | CR | W |
R4 | Fuel 1 | W | W | CR | CR |
R5 | Fuel 1 | CR | CR | W | CR |
R6 | Fuel 1 | CR | CR | CR | CR |
R7 | Fuel 2 |
In the two-dimensional case, we use FreeFEM to solve the problems listed in Table 5. We chose uniform grids with
Using GIPMNN and PC-GIMPNN to Solve for Eigenvalue
The number of points N=NxNy is used to calculate Lossgipmnn and number of points
The 2D reactor is divided into six parts, as shown in Figure 7. The output layer includes six neurons, and there are six functions that are defined in different subdomains and are labeled as u, v, w, r, p, and q, where u, v, w, and r denote functions defined in CR1, CR2, CR3, and CR4, and p and q denote functions defined for Fuel and Graphite, respectively. SCR1, SCR2, SCR3, SCR4, and SGF denote different datasets at different interfaces.
For PC-GIPMNN, interface conditions (29) and (30) are implemented as loss functions (43) and (44) in the 2D example.
The eigenfunction is expressed as follows:
Using DRM to Solve for Eigenvalue and the failure of DRM during the 2D Experiments
Given that the homogeneous Neumann boundary condition is used, the loss function in DRM can be defined as:
We use the number of epochs Nepoch=500000 in DRM. First, we found that the DRM can learn the eigenvalues and eigenfunctions at an early stage of the training process. Subsequently, the DRM attempts to learn a smaller eigenvalue after it is close to the true eigenvalue. Finally, the DRM successfully learns one smaller eigenvalue that may be close to the true eigenvalue and one terrible eigenfunction that is meaningless and far from the true eigenfunction. This phenomenon is illustrated in Fig. 8.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F008.jpg)
As listed in Table 6, although the neural network in DRM fails to learn the eigenfunction, the eigenvalue is close to the true value. This phenomenon may be caused by minimization of Rayleigh quotient. As mentioned in [49], the loss approaches zero, whereas the eigenvalue may not reach zero.
In Figure 8, we observe that the failure of DRM during the 2D experiments occurs after
Results
Similar to the results of the one-dimensional case, the relative errors of keff and eigenfunction ϕ are shown in Tables 7 and 8. It can be observed that both the relative errors for keff obtained via all the methods are small, and the results for the DRM are trained with Nepoch=50000.
Problem | keff (FEM) | keff (GIPMNN) | keff (PC-GIPMNN) | keff (DRM) | |||
---|---|---|---|---|---|---|---|
F1 | 1.0118 | 1.0094 | 1.0121 | 1.0096 | 2.3574×10-03 | 2.8227×10-04 | 2.1558×10-03 |
F2 | 1.0052 | 1.0023 | 1.0055 | 1.0024 | 2.8782×10-03 | 2.9864×10-04 | 2.7674×10-03 |
F3 | 1.0000 | 0.9968 | 0.9999 | 0.9965 | 3.2380×10-03 | 1.2475×10-04 | 3.4253×10-03 |
F4 | 1.0079 | 1.0053 | 1.0081 | 1.0054 | 2.5913×10-03 | 2.4834×10-04 | 2.4439×10-03 |
F5 | 1.0052 | 1.0025 | 1.0054 | 1.0024 | 2.7120×10-03 | 3.1435×10-04 | 2.7394×10-03 |
R1 | 1.0118 | 1.0094 | 1.0121 | 1.0096 | 2.3574×10-03 | 2.8227×10-04 | 2.1558×10-03 |
R2 | 1.0052 | 1.0023 | 1.0055 | 1.0024 | 2.8782×10-03 | 2.9864×10-04 | 2.7674×10-03 |
R3 | 0.9921 | 0.9937 | 0.9927 | 0.9878 | 1.5981×10-03 | 6.0649×10-04 | 4.3297×10-03 |
R4 | 1.0017 | 1.0014 | 1.0015 | 0.9987 | 3.5502×10-04 | 2.4851×10-04 | 3.0409×10-03 |
R5 | 0.9780 | 0.9759 | 0.9781 | 0.9721 | 2.1223×10-03 | 1.6160×10-04 | 6.0148×10-03 |
R6 | 0.9668 | 0.9639 | 0.9680 | 0.9584 | 3.0188×10-03 | 1.2722×10-03 | 8.6348×10-03 |
R7 | 1.1018 | 1.0929 | 1.1020 | 1.0953 | 8.0309×10-03 | 2.1827×10-04 | 5.9042×10-03 |
Problem | ϕrel (GIPMNN) | ϕrel (PC-GIPMNN) | ϕrel (DRM) |
---|---|---|---|
F1 | 4.6050×10-02 | 3.0937×10-02 | 4.4186×10-02 |
F2 | 8.9731×10-02 | 3.6713×10-02 | 7.0763×10-02 |
F3 | 8.9385×10-02 | 4.1200×10-02 | 6.7164×10-02 |
F4 | 8.5110×10-02 | 4.3430×10-02 | 5.9907×10-02 |
F5 | 7.9413×10-02 | 4.6256×10-02 | 6.6049×10-02 |
R1 | 4.6050×10-02 | 3.0937×10-02 | 4.4186×10-02 |
R2 | 8.9731×10-02 | 3.6713×10-02 | 7.0763×10-02 |
R3 | 1.1047×10-01 | 6.2990×10-02 | 8.4720×10-02 |
R4 | 1.3013×10-01 | 2.9511×10-02 | 6.8647×10-02 |
R5 | 2.5852×10-01 | 2.9718×10-02 | 7.7923×10-02 |
R6 | 4.6253×10-01 | 5.1511×10-02 | 9.1424×10-02 |
R7 | 1.1612×10-01 | 1.8635×10-02 | 7.2148×10-02 |
For problems F1, F2, F4, and R7, the relative error of keff obtained by DRM was smaller than that obtained by GIPMNN, except for problems F3, F5, R3, R4, R5, and R6. However, although the relative error of keff obtained by GIPMNN is small, the relative error of ϕ simulated by GIPMNN is larger than that obtained by DRM. It is observed that
In Fig. 9 and 10, the eigenfunctions computed by the FEM are shown in the first column, and the relative errors of the eigenfunctions obtained by the GIPMNN, PC-GIPMNN, and DRM are shown in the other columns for different problems. It is observed that the relative errors of the eigenfunction computed by the PC-GIPMNN and DRM are smaller than those obtained by the GIPMNN, which failed to learn some details. Compared to the eigenfunctions computed by the FEM, the results obtained by the PC-GIPMNN are the best among the three methods.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F009.jpg)
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F010.jpg)
2D IAEA Benchmark Problem
We also considered the classical 2D IAEA benchmark problem reported in the study by Yang et al. [40], which was modeled using two-dimensional and two-group diffusion equations. Here, one-group neutron diffusion equation, defined in Eq. (4)m is considered, and multigroup problems are considered in our future study. Its geometry is shown in Fig. 1. The main bulk of the reactor consists of two fuel regions, labeled 1 and 2, representing the two types of fuel materials. Within its central region, there are four control rods, which are all labeled as 3. The last region, labeled 4, was composed of water. Cross-sectional data for the 2D IAEA benchmark problem are presented in Table 9. It is worth noting that only one quarter of the reactor is shown in this figure because the rest can be inferred by symmetry along the x- and y-axes. Therefore, this 2D IAEA benchmark problem is confined to the two types of boundary conditions defined in Eq. (7). The problem is confined to the Neumann boundary condition on the x- and y-axes and to the Robin boundary condition on the other boundaries.
Region | ∑a(cm-1) | ∑s(cm-1) | v ∑f(cm-1) | Material |
---|---|---|---|---|
1 | 2.0 | 0.53 | 0.079 | Fuel 1 |
2 | 0.087 | 0.55 | 0.085 | Fuel 2 |
3 | 0.38 | 0.20 | 0.0 | Control rod |
4 | 0.01 | 0.89 | 0.0 | Water |
In this two-dimensional case, we used FreeFEM to solve the 2D IAEA benchmark problem with the parameters listed in Table 9. We selected uniform grids with
Using GIPMNN and PC-GIMPNN to Solve for Eigenvalue
The number of points N=NxNy is used to calculate Lossgipmnn. The number of points NNb on the x- and y-axes and number of points NRb on the other boundaries were used to calculate LossNb and LossRb, which enforced the Neumann and Robin boundary conditions. The number of neurons was 20 for each hidden layer in ResNet and Nepoch=500000 for GIPMNN and PC-GIPMNN. As mentioned previously, we used the optimal ratio of the outer and inner iterations, which was 1:1.
The 2D reactor is divided into seven parts, as shown in Fig. 11. The output layer has seven neurons, and seven functions are defined in different subdomains and labeled as u, v, w, r, p, q, and h, where u, v, w, and r denote the functions defined in the control rods and p, q, and h are the functions defined in the fuel and water regions, respectively. Sup, Svp, Swp, Srp, Srq, Spq, and Sqh denote different datasets at different interfaces.
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F011.jpg)
For the PC-GIPMNN, the interface conditions (29) and (30) are implemented as the loss functions (47) and (48), respectively, in the 2D example.
The eigenfunction can be expressed as follows:
Using DRM to Solve for Eigenvalue
Given that the homogeneous Neumann boundary condition is used, the loss function in DRM omits the impact of the Neumann boundary condition and focuses on the Robin boundary condition. The loss function is defined as follows:
Results
As discussed above, we also trained the DRM with the number of epochs Nepoch=500000 and found that DRM failed to learn the eigenfunction again. In this case, DRM attains a good keff=0.9750 and the relative errors of keff and ϕ are 6.4023E-03 and 0.9557, respectively. Therefore, we retrained the DRM with Nepoch=50000 and stored the best results.
The relative errors of keff and eigenfunction ϕ are shown in Table 10 and Table 11. It was observed that all three methods obtained good results, and the relative errors of keff of DRM were small. However, the ability of DRM to learn the eigenfunction was worse than that of GIPMNN and PC-GIPMNN and the relative errors of ϕ were the largest. Thus, it can be concluded that the PC-GIPMNN successfully learned the eigenvalues and eigenfunctions. The same conclusion can be drawn from the graphs in Fig. 12.
Problem | keff (FEM) | keff (GIPMNN) | keff (PC-GIPMNN) | keff (DRM) | |||
---|---|---|---|---|---|---|---|
IAEA | 0.9688 | 0.9692 | 0.9691 | 0.9685 | 4.0370×10-04 | 3.0812×10-04 | 2.8218×10-04 |
Problem | ϕrel (GIPMNN) | ϕrel (PC-GIPMNN) | ϕrel (DRM) |
---|---|---|---|
IAEA | 8.5688×10-02 | 4.7381×10-02 | 8.9602×10-02 |
-202310/1001-8042-34-10-017/alternativeImage/1001-8042-34-10-017-F012.jpg)
For the 1D slab reactor, the computation time of FEM is 1.25 s, and the training times of DRM, GIPMNN, and PC-GIPMNN are 4788.37 s, 10614.57 s, and 16052.16 s, respectively. For the 2D reactor, the computation time of FEM is 3.64 s, and the training times of DRM, GIPMNN, and PC-GIPMNN are 5827.91 s, 18444.39 s, and 108072.41 s, respectively. For the 2D IAEA benchmark, the computation time of FEM is 5.22 s, and the training times of DRM, GIPMNN, and PC-GIPMNN are 29352.83 s, 64546.74 s, and 137812.92 s, respectively.
Although PC-GIPMNN was better than DRM and GIPMNN, it required significantly more training time to obtain accurate results. Compared with FEM, neural network methods require excessive time. However, neural networks are an emerging method and we believe that they will achieve better results in the near future.
Conclusion
In this study, we proposed two methods, GIPMNN and PC-GIPMNN, to solve generalized K-eigenvalue problems in nuclear reactor physics. We also conducted a comprehensive study of GIPMNN, PC-GIPMNN, and DRM. GIPMNN follows the main idea of the inverse power method to find the smallest eigenvalue. The PC-GIPMNN enforce interface conditions through multiple neurons in the output layer. The concept of DRM is to define the function of the Rayleigh quotient and form an optimization problem. Unlike DRM solving for the smallest eigenvalue by directly minimizing the eigenvalue (Rayleigh quotient), the GIPMNN and PC-GIPMNN attain the smallest eigenvalue using the iterative method. All the methods used neural networks to represent functions and the differential was implemented using AD. Finally, we applied these three methods to problems in reactor physics.
Three numerical experiments were conducted to verify the applicability and accuracy of the GIPMNN, PC-GIPMNN, and DRM. In the first 1D example, we used inner and outer iterations for the simulation. According to our test, the best ratio of outer and inner iterations was 1:1. Furthermore, we compared the results of the GIPMNN, PC-GIPMNN, and DRM with those of the FEM. For the continuous problem, the solution learned by the GIPMNN was more accurate than those learned by the DRM and PC-GIPMNN. For interface problem, the eigenvalue and eigenfunction learned by PC-GIPMNN were better than that learned by DRM and GIPMNN. This is due to the interface conditions that are implemented in the loss function of PC-GIPMNN.
In the 2D examples, we observed the failure of DRM on the 2D experiments. The DRM can learn the eigenvalue and eigenfunction at the early stage of the training process, and the results decrease when Nepoch increases. Therefore, we selected Nepoch=50000 to train the DRM and compare the results obtained by the GIPMNN, PC-GIPMNN, and DRM with those obtained by the FEM. The results show that the PC-GIPMNN outperforms the GIPMNN and DRM for the results of eigenvalue and eigenfunction. Moreover, the GIPMNN and PC-GIPMNN are more stable than the DRM.
Although good results were obtained, there are still some aspects that need to be examined in the future. First, given that the architecture of a neural network significantly influences the accuracy of our methods, it is important to design a universal network architecture to achieve high accuracy. For example, MLP is widely used in the current field of solving PDEs using neural networks, and ResNet [60] is effective in improving the convergence rate and may even obtain better results than MLP. Recently, a transformer [74] was used for operator learning and better results were obtained. Sifan Wang et al. [62] investigated the effect of MLP on operator learning and proposed a modified MLP [61], which is a new architecture of MLP that improves accuracy. Although GIPMNN and PC-GIPMNN are more stable than DRM and PC-GIPMNN is more accurate than DRM, they require a large number of iterations and a long training time to achieve good results. Therefore, improving convergence and reducing training time will be investigated in our future study. Furthermore, the failure of DRM on 2D experiments on the eigenvalue problem should be studied and clarified. Finally, for the interface problem, a suitable sampling algorithm can facilitate the training process and provide a better approximation. The next study could also involve the implementation of the proposed networks in emerging reactor digital twins [75-77] as core solvers.
Nuclear reactor core analysis code: Citation
. (Development of the nodal collocation method for solving the neutron diffusion equation
. Annals of Nuclear Energy 14, 527-541 (1987). doi: 10.1016/0306-4549(87)90074-0The application of the finite element method to the multigroup neutron diffusion equation
. Nuclear Science and Engineering 47, 302-310 (1972). doi: 10.13182/NSE72-A22416Application of a dual variational formulation to finite element reactor calculations
. Annals of Nuclear Energy 20, 823-845 (1993). doi: 10.1016/0306-4549(93)90076-2Solution of neutron diffusion problems by discontinuous galerkin finite element method with consideration of discontinuity factors
. Journal of Nuclear Engineering and Radiation Science 9, 031503 (2023). doi: 10.1115/1.4055379Progress in nodal methods for the solution of the neutron diffusion and transport equations
. Progress in Nuclear Energy 17, 271-301 (1986). doi: 10.1016/0149-1970(86)90034-XAn analytic nodal method for solving the two-group, multidimensional, static and transient neutron diusion equation
. (Development and validation of reactor nuclear design code corca-3d
. Nuclear Engineering and Technology 51, 1721-1728 (2019). doi: 10.1016/j.net.2019.05.015Iterative methods for solving nonlinear problems of nuclear reactor criticality
. Physics of Atomic Nuclei 75, 1551-1556 (2012). doi: 10.1134/S1063778812130042Das-pinns: A deep adaptive sampling method for solving high-dimensional partial differential equations
. Journal of Computational Physics 476, 111868 (2023). doi: 10.1016/j.jcp.2022.111868An inverse-distance-based fitting term for 3d-var data assimilation in nuclear core simulation
. Annals of Nuclear Energy 141, 107346 (2020). doi: 10.1016/j.anucene.2020.107346Observation data compression for variational assimilation of dynamical systems
. Journal of Computational Science 53, 101405 (2021). doi: 10.1016/j.jocs.2021.101405Hybrid data assimilation methods, part i: Numerical comparison between geim and pbdw
. Annals of Nuclear Energy 190, 109864 (2023). doi: 10.1016/j.anucene.2023.109864Hybrid data assimilation methods, part ii: Application to the dynasty experimental facility
. Annals of Nuclear Energy 190, 109863 (2023). doi: 10.1016/j.anucene.2023.109863Partial differential equations
. (2010) https://bookstore.ams.org/gsm-19-r/Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations
. Communications in mathematics and statistics 5, 349-380 (2017). doi: 10.1007/s40304-017-0117-6Solving high-dimensional partial differential equations using deep learning
. Proceedings of the National Academy of Sciences 115, 8505-8510 (2018). doi: 10.1073/pnas.1718942115Dgm: A deep learning algorithm for solving partial differential equations
. Journal of computational physics 375, 1339-1364 (2018). doi: 10.1016/j.jcp.2018.08.029The deep ritz method: a deep learning-based numerical algorithm for solving variational problems
. Communications in Mathematics and Statistics 6, 1-12 (2018). doi: 10.1007/s40304-018-0127-zPhysics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations
. Journal of Computational physics 378, 686-707 (2019). doi: 10.1016/j.jcp.2018.10.045Physics-informed neural network for nonlinear dynamics in fiber optics
. Laser & Photonics Reviews 16, 2100483 (2022). doi: 10.1002/lpor.202100483Enhanced physics-informed neural networks for hyperelasticity
. International Journal for Numerical Methods in Engineering 124, 1585-1601 (2023). doi: 10.1002/nme.7176Meshless physics-informed deep learning method for three-dimensional solid mechanics
. International Journal for Numerical Methods in Engineering 122, 7182-7201 (2021). doi: 10.1002/nme.6828Simulation of multi-species flow and heat transfer using physics-informed neural networks
. Physics of Fluids 33, 087101 (2021). doi: 10.1063/5.0058529Physics-informed neural networks for heat transfer problems
. Journal of Heat Transfer 143, 060801 (2021). doi: 10.1115/1.4050542Data-driven modeling of coarse mesh turbulence for reactor transient analysis using convolutional recurrent neural networks
. Nuclear Engineering and Design 390, 111716 (2022). doi: 10.1016/j.nucengdes.2022.111716Physics-informed neural networks for inverse problems in nano-optics and metamaterials
. Optics express 28, 11618-11633 (2020). doi: 10.1364/OE.384875Data-driven surrogate model with latent data assimilation: Application to wildfire forecasting
. Journal of Computational Physics 464, 111302 (2022). doi: 10.1016/j.jcp.2022.111302Generalised latent assimilation in heterogeneous reduced spaces with machine learning surrogate models
. Journal of Scientific Computing 94, 11 (2023). doi: 10.1007/s10915-022-02059-4Wasserstein generative adversarial uncertainty quantification in physics-informed neural networks
. Journal of Computational Physics 463, 111270 (2022). doi: 10.1016/j.jcp.2022.111270Uncertainty quantification in estimating blood alcohol concentration from transdermal alcohol level with physics-informed neural networks
.Adversarial uncertainty quantification in physics-informed neural networks
. Journal of Computational Physics 394, 136-152 (2019). doi: 10.1016/j.jcp.2019.05.027Uncertainty quantification for multiphase-cfd simulations of bubbly flows: a machine learning-based bayesian approach supported by high-resolution experiments
. Reliability Engineering & System Safety 212, 107636 (2021). doi: 10.1016/j.ress.2021.107636A data-enabled physics-informed neural network with comprehensive numerical study on solving neutron diffusion eigenvalue problems
. Annals of Nuclear Energy 183, 109656 (2023). doi: 10.1016/j.anucene.2022.109656Deep learning solution of the eigenvalue problem for differential operators
. Neural Computation 35, 1100-1134 (2023). doi: 10.1162/neco_a_01583Physics-informed neural networks for quantum eigenvalue problems
.Physics-informed neural network method and application to nuclear reactor calculations: A pilot study
. Nuclear Science and Engineering 197, 1-22 (2022). doi: 10.1080/00295639.2022.2123211Variational principles
. Variational Principles of Continuum Mechanics 3–44 (2009). doi: 10.1007/978-3-540-88467-5_1Solving high-dimensional eigenvalue problems using deep neural networks: A diffusion monte carlo like approach
. Journal of Computational Physics 423, 109792 (2020). doi: 10.1016/j.jcp.2020.109792Neural networks base on power method and inverse power method for solving linear eigenvalue problems
. Computers & Mathematics with Applications 147, 14-24 (2023). doi: 10.1016/j.camwa.2023.07.013Extension of the pinn diffusion model to k-eigenvalue problems
. (2022) https://inis.iaea.org/search/search.aspx?orig_q=RN:54002255Conservative physics-informed neural networks on discrete domains for conservation laws: Applications to forward and inverse problems
. Computer Methods in Applied Mechanics and Engineering 365, 113028 (2020). doi: 10.1016/j.cma.2020.113028Surrogate modeling for neutron diffusion problems based on conservative physics-informed neural networks with boundary conditions enforcement
. Annals of Nuclear Energy 176, 109234 (2022). doi: 10.1016/j.anucene.2022.109234An autoencoder-based reduced-order model for eigenvalue problems with application to neutron diffusion
. International Journal for Numerical Methods in Engineering 122, 3780-3811 (2021). doi: 10.1002/nme.6681A mesh-free method for interface problems using the deep learning approach
. Journal of Computational Physics 400, 108963 (2019). doi: 10.1016/j.jcp.2019.108963A mesh-free method using piecewise deep neural network for elliptic interface problems
. Journal of Computational and Applied Mathematics 412, 114358 (2022). doi: 10.1016/j.cam.2022.114358A discontinuity capturing shallow neural network for elliptic interface problems
. Journal of Computational Physics 469, 111576 (2022). doi: 10.1016/j.jcp.2022.111576Deep residual learning for image recognition
. (Understanding and mitigating gradient flow pathologies in physics-informed neural networks
. SIAM Journal on Scientific Computing 43, A3055-A3081 (2021). doi: 10.1137/20M1318043Improved architectures and training algorithms for deep operator networks
. Journal of Scientific Computing 92, 35 (2022). doi: 10.1007/s10915-022-01881-0A method for representing periodic functions and enforcing exactly periodic boundary conditions with deep neural networks
. Journal of Computational Physics 435, 110242 (2021). doi: 10.1016/j.jcp.2021.110242A pod reduced-order model for eigenvalue problems with application to reactor physics
. International Journal for Numerical Methods in Engineering 95, 1011-1032 (2013). doi: 10.1002/nme.4533New development in freefem++
. Journal of Numerical Mathematics 20, 251-266 (2012). doi: 10.1515/jnum-2012-0013Sensor placement in nuclear reactors based on the generalized empirical interpolation method
. Journal of Computational Physics 363, 354-370 (2018). doi: 10.1016/j.jcp.2018.02.050Reactor power distribution detection and estimation via a stabilized gappy proper orthogonal decomposition method
. Nuclear Engineering and Design 370, 110833 (2020). doi: 10.1016/j.nucengdes.2020.110833Neutron noise calculation: A comparative study between sp3 theory and diffusion theory
. Annals of Nuclear Energy 156, 108184 (2021). doi: 10.1016/j.anucene.2021.108184Optimal and fast field reconstruction with reduced basis and limited observations: Application to reactor core online monitoring
. Nuclear Engineering and Design 377, 111113 (2021). doi: 10.1016/j.nucengdes.2021.111113Data-enabled physics-informed machine learning for reduced-order modeling digital twin: application to nuclear reactor physics
. Nuclear Science and Engineering 196, 668-693 (2022). doi: 10.1080/00295639.2021.2014752Choose a transformer: Fourier or galerkin
. Advances in neural information processing systems 34, 24924-24940 (2021).Parameter identification and state estimation for nuclear reactor operation digital twin
. Annals of Nuclear Energy 180, 109497 (2023). doi: 10.1016/j.anucene.2022.109497An efficient digital twin based on machine learning svd autoencoder and generalised latent assimilation for nuclear reactor physics
. Annals of Nuclear Energy 179, 109431 (2022). doi: 10.1016/j.anucene.2022.109431Data-enabled physics-informed machine learning for reduced-order modeling digital twin: Application to nuclear reactor physics
. Nuclear Science and Engineering 196, 668-693 (2022). doi: 10.1080/00295639.2021.2014752The authors declare no conflict of interest.