Introduction
High-fidelity neutron-shielding calculations are an essential part of advanced nuclear reactor system design. The discrete ordinate (SN) method is one of the most popular deterministic particle transport methods owing to its high angular resolution [1]. Reliable modeling and simulation of SN transport depend on two key essential conditions and fundamental components: reasonable spatial discretization schemes and mesh distributions with sufficient modeling fidelity. The two key factors interacted with and supported each other throughout the process.
Conventional Cartesian structured grids are widely used and implemented because of their regular arrangement and friendliness to finite-difference (FD) spatial discretization schemes [2-4]. However, there is no general robust meshing algorithm for acquiring exact descriptions of complex structures for realistic modeling and simulations [5]. Adaptive mesh refinement techniques retain the original mesh distribution features and mark cells with large errors for refinement [6]. Local error estimation techniques and geometric data structures play significant roles in determining the stabilities and computational efficiencies of these systems. Unstructured grids use piecewise flat or curved surfaces to provide complicated geometries with the highest possible flexibility. With the development of the finite element method (FEM), unstructured grids have been widely applied to reactor physical analyses [7], multiphysics simulations [8], and thermal-fluid dynamics [9]. Combining finite element numerical methods and the SN method with unstructured meshes has become a trend in nuclear shielding design.
Unfortunately, the first-order SN equation cannot be solved stably using the continuous FEM [10]. However, while the solutions to this hyperbolic equation are smooth along the streamlines, they may be discontinuous perpendicular to the chosen direction. To obtain a robust solution, several Petrov-Galerkin (PG) approaches that update the test function as the sum of the trial functions and introduce an additional stabilization term [11-13]. It was observed that the free stabilization parameter affects the convergence of the SN equation and raises uncertainty because it is usually considered as a matter of experience. Additionally, PGs cannot generally produce symmetric positive-definite coefficient matrices or inherit the best approximation property when the transport equations contain nonself-adjoint operators.
The discontinuous Galerkin finite element method (DGFEM) is a different category of stability improvement. This was pioneered by Reed and Hill for neutron transport problems in the early 1970s [14]. The DGFEM connects adjacent elements through the incoming interface fluxes from the upwind (upstream) meshes and applies the usual FEM procedure to a single mesh. However, the number of unknowns in the DGFEM is much larger than that in the continuous FEMs, which demands a matrix assembly and solution method. The standard matrix assembly procedure looks for a reliable iterative solver for a banded sparse system matrix solution and must handle the storage brought on by the increased freedom [15]. The matrix-free strategy stores the diagonal blocks and realizes the inversion of the transport operator by means of the matrix-vector product; however, a sum factorization technique to enable translation between vector entries and values or gradients in quadrature points is required [16]. Another commonly used technique is the conventional sweeping procedure, such as FDs, in which parallelization of the sweeping procedure is implemented to reduce the computing time, which may become important in a large SN order.
In terms of parallelization, complex grid connection relationships, and irregular element sweep orders pose challenges to the algorithms. In 2000, Plimpton described an asynchronous, parallel message-passing algorithm that simultaneously performs sweeps from many directions across unstructured grids [17]. In 2002, Pautz used a low-complexity list-ordering heuristic to determine sweep ordering on a partitioned mesh [18]. The essence of these two algorithms is the conversion of the abstract element sweep order into a directed acyclic graph (DAG). Since 2010, several research institutions have conducted extensive studies on high-performance architectural applications [19, 20].
A multigroup 2D SN transport code, ThorSNIPE, with unstructured meshes was developed and validated [21]. This code was extended to 3D SN transport calculations and employed a multithreaded parallel upwind sweep algorithm to achieve a more accurate description and improve the reliability and efficiency of transport calculations in complex geometries. The finite element method was implemented using an open-source finite element library deal.II-9.4.0 [22]. The deal.II is an open-source C++ program library targeted at the computational solution of partial differential equations using adaptive finite elements and focuses on generality, dimension-independent programming, parallelism, and extensibility. It includes many state-of-the-art programming techniques for solving partial differential equations, linear algebra problems, and computer science strategies and offers users a modern interface for complex data structures and algorithms. With the assistance of this library, ThorSNIPE can be made dimension-independent, supporting different mesh sizes and types of finite elements.
The remainder of this paper is organized as follows. In Sect. 2, the main theory and methodology of the ThorSNIPE code are described in detail, including the SN weak-form equation based on DGFEM theory and the multithreaded parallel upwind sweep algorithm. The numerical results are presented and discussed in Sect. 3. The conclusions are summarized in Sect. 4.
Theory and methodology
Weak form of the discrete ordinates method equations
The SN form was obtained by solving the Boltzmann transport equation along discrete directions and replacing the integrals over the unit sphere with quadrature sets. The steady-state multi-group SN neutron transport equation can be written as
Eq. (1) is a linear hyperbolic equation: It is necessary to transform this "strong form" differential-integral equation into "weak form" for solving. The spatial domain
The divergence theorem is then applied to Eq. (5), yielding:
Substituting Eq. (7) and (8) into (6),
The general boundary can be divided into three types based on its relative position: interior boundary, outflow domain boundary
The discontinuous Galerkin finite element discretization scheme
The Galerkin method is theoretically based on the weighted residual method, which multiplies the original partial differential equation by a test function. Seeking a numerical approximation of the angular flux term using Lagrange polynomials of degree P
In Galerkin discretization schemes, test functions share similar function space as basis functions,
Substituting Eq. (12) and (13) into (10),
The upwind scheme is the essence of the DGFEM. Notably, although the test functions on the edges are always taken from within the element, the primal functions on the edges are taken upwind with respect to the streaming direction. For outgoing edges, the primal functions are taken from within the element itself, whereas for incoming edges, the primal functions are taken from the upwind neighbor element.
Multithreaded parallel upwind sweep algorithm
Cell-based sweep algorithm
In the SN method, a transport sweep was performed for all elements along each discrete direction. The ordering in which the elements and degrees of freedom (DOF) traverse affects the speed of convergence. As mentioned in Section 1, the DGFEM makes it possible to use a cell-based sweeping procedure in SN methods rather than assembling a global matrix.
The element sweep order for each discrete direction in the ThorSNIPE code was developed using an upwind scheme based on the deal.II library. In our case, an array with all relevant cells is created and sorted in the upwind direction using a "dealii :: DoFRenumbering :: CompareDownstream" object in the deal.II FEM library. A straightforward 2D model with three material regions and various sweep schemes is shown in Fig. 1. Cells in a conventional scheme are often categorized from edge to center to make it easier to facilitate the creation of a global matrix. The downwind system performs renumbering in the direction opposite to the discrete direction, whereas the upwind scheme manages renumbering alongside it. The definitions of "upwind" and "downwind" are determined by multiplying the unit normal vector to the cell's surface by the cosines of the corresponding discrete direction. In reality, the mesh location, discrete direction, and sweep-define the order of each element. The sweep solution technique also complies with neutron transport laws. Not all pathways are mutually exclusive, and each distinct direction results in a unique order in which this solution technique can be applied.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F001.jpg)
From the perspective of solving the equations, various sweep schemes also affect the calculation of the incoming angular fluxes
Algorithm 1 displays the cell-based sweep algorithm for each group and angle once the sweep order has been determined. The local equation can be divided into two parts as follows: The transport matrix first makes contributions from the streaming and removal operators determined by the fixed mass matrix and total cross-section. Second, the right-hand side (RHS) vector includes contributions from scattering source operations and fixed source or fission source operators. The in-group and cross-group couplings comprised the contributions of the scattering sources. Lastly, a "dealii :: types :: global dof index" object translates the local solution to the global angular solution. This cell-based sweep technique delivers the DGFEM solution to all cells in a single sweep order without coupling integral terms, fully utilizing the key benefits of DGFEM approaches.
Input: g, m, qfixed(r) | ||||||
for upwind sorted cell do | ||||||
for DOF of cell i do | ||||||
for DOF of cell |
||||||
for moment n do | ||||||
for moment k do | ||||||
for DOF of cell i do | ||||||
for faces on cell do | ||||||
if outflow face then | ||||||
for DOF of face i do | ||||||
for DOF of face j do | ||||||
else if inflow face then | ||||||
for DOF of face i do | ||||||
cell_rhs(i)=source_rhs(i)+boundary_rhs(i) | ||||||
cell_matrix(i,j)=transport_matrix(i,j)+boundary_matrix(i,j) | ||||||
solution(i)=cell_matrix(i,j)-1cell_rhs(i) | ||||||
for DOF of cell i do | ||||||
angular_solution[m](local_dof_indices[i])=solution(i) |
The following two steps were employed to compute the local contributions of an individual cell and RHS: First, the integral is transformed from an actual cell
Multithread parallel strategy
Parallel transport computation has gained popularity in the nuclear field because of the emergence of high-performance computing clusters [24]. How the processors schedule parallel jobs significantly affects the performance of code using parallel computing. The two basic techniques are spatial-domain decomposition (SDD) [25] and angular-domain decomposition (ADD) [26]. SDD techniques subdivide the spatial domain into subdomains, such that the common mesh-sweep method is performed concurrently in numerous subdomains at diverse angles. However, parallelizing the SN spatial sweep is more challenging because of the upwind-downwind dependencies between the domain boundaries.
ADD techniques are frequently considered more expandable and require less memory [27]. They entail a mesh sweep along a similar quadrant and are the same for all angles in the chosen quadrature [28]. All operations are statically scheduled to the participating processors without negatively affecting the load balance and are ideally suited for personal computers (PCs). In shared memory machines, the traditional approach to parallelism is to break code down into threads. The deal.II leverages the Threading Building Blocks (TBB) library [29] as basic wrappers to build an object called "tasks.” TBB offers transferable interfaces across many different platforms and abstracts the specifics of run-state signals onto individual threads.
The entire procedure for the multithreaded parallel approach in the ThorSNIPE code is shown in Fig. 2. Both inner and outer iterations contain an angular loop. The number of CPU cores determines the total number of tasks in a PC system. The angular loops were divided into a specific number of subranges that were evenly distributed across the threads. A cell-based sweep sequence is used for each angular loop. A subrange is a task. The tasks are essentially independent during the iterations. They performed the calculations using the same mesh and equation structure as the respective quadrature set. However, the storage of angular fluxes is based on the local DOF indices calculated using sweep schemes. The calculation results for different subranges were collected to update the scattering source. Following the execution of the subranges, threads keep themselves busy by stealing complete portions of the subranges from other threads if they have finished their work. Thus, the code can fully exploit the system thread resource, which requires frequent interruptions by the operating system to allow other threads to execute on the available processor cores.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F002.jpg)
Results and discussions
The multithreaded parallel upwind sweep algorithm is validated using three typical benchmark problems. The IAEA benchmark with five different regions was used to test the performances of the different sweep schemes. The Kobayashi-3i benchmark with cavities and bend ducts was adopted to evaluate the performance of the parallel computation for different mesh sizes and angular quadrate sets. The VENUS-3 benchmark was used to validate the algorithm’s capabilities for realistic transport calculations. In all benchmarks, convergence criteria are 10-5 for the eigenvalue and 10-4 for the average flux in each space interval.
IAEA benchmark
The IAEA benchmark is a 2D five-region calculation problem consisting of two large source zones and two large absorber zones surrounded by light water [30] and is designed for IAEA advanced reactor neutron transport computation. There is a strong nonuniformity in the fuel region, as shown in Fig. 3. The cross-sections are given in Table 1, and all boundary conditions are vacuum.
Region | ||||
---|---|---|---|---|
R1 | 0.60 | 0.079 | 0.53 | 1.0 |
R2 | 0.48 | 0.0 | 0.20 | 0.0 |
R3 | 0.70 | 0.0 | 0.66 | 1.0 |
R4 | 0.65 | 0.043 | 0.50 | 0.0 |
R5 | 0.90 | 0.0 | 0.83 | 0.0 |
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F003.jpg)
This benchmark problem is mainly used to analyze the effect of the element sweep order on the cell-based algorithm efficiency and assert the superiority of the upwind sweep scheme. Under the PNTN-S8 angular approximation, comparisons of the numerical results for the fixed source and eigenvalue cases simulated with 2338 elements are presented in Tables 2 and 3, respectively. The relative deviations in the average flux were within 2.50% in the fixed-source case and within 5.00% in the eigenvalue case. All the numerical results show excellent agreement with the references, demonstrating the accuracy and validity of the algorithm and modeling approach for such strongly heterogeneous problems. The change in the element sweep order does not make a large difference to the results. This is because the transport sweep operations were performed individually on each element, and the results were added. Because matrix addition is commutative and these operators in the equation can be described as applying a sum of matrices, the orders of various components will not impact the final result.
Reference | Average flux (Relative errors) cm -2 s -1 | |||
---|---|---|---|---|
Conventional | Downwind | Upwind | ||
R1 | 11.941 | 11.872 (0.58%) | 11.874 (0.56%) | 11.871 (0.59%) |
R2 | 0.545 | 0.547 (0.37%) | 0.547 (0.37%) | 0.546 (0.18%) |
R3 | 19.168 | 19.045 (0.64%) | 19.051 (0.61%) | 19.039 (0.67%) |
R4 | 0.843 | 0.838 (0.60%) | 0.840 (0.36%) | 0.837 (0.71%) |
R5 | 1.530 | 1.542 (0.78%) | 1.553 (1.50%) | 1.531 (0.07%) |
Reference | Average flux (Relative Errors) cm -2 s -1 | |||
---|---|---|---|---|
Conventional | Downwind | Upwind | ||
R1 | 0.01686 | 0.016371(2.90%) | 0.01661(1.18%) | 0.016643(1.29%) |
R2 | 0.000125 | 0.000131(4.80%) | 0.000128(2.40%) | 0.000128(2.40%) |
R3 | 0.000041 | 0.000040(2.44%) | 0.000041(0.00%) | 0.000041(0.00%) |
R4 | 0.000295 | 0.000295(0.00%) | 0.000293(0.68%) | 0.000293(0.68%) |
R5 | 0.000791 | 0.000799(1.01%) | 0.000788(0.38%) | 0.000789(0.25%) |
keff | 1.0083 | 1.007076 | 1.008371 | 1.008305 |
pcm | -12 | 7 | 1 |
As shown in Fig. 4, the number of iterations was compared with the change in the number of elements in the fixed-source calculation. The numerical results indicated that the upwind sweep scheme had a stable convergence speed, regardless of the number of elements. The relationship between the number of iterations and elements grows exponentially in both types of sweep schemes. The primary functions on the edges are determined based on the upwind edges with regard to the streaming direction, and transport problems involve a directional flow of information. The upwind scheme fully uses the initial boundary conditions and provides spatial flux moments for the neighboring elements. The other sweep schemes increased the number of source and power iterations required to meet the convergence criteria in the subsequent matrix calculations to compensate for the lack of constraints. In addition, no overhead cost of assembling and solving the stiffness matrix in the current framework was observed for the three sweep schemes mentioned above. Therefore, the upwind sweep algorithm can save significant computational costs with fewer iterations and is suitable for DGFEM simulations.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F004.jpg)
Kobayashi-3i benchmark
The Kobayashi-3i benchmark [31] is a bend duct-type problem with void regions in a highly absorbing medium, as shown in Fig. 5. The cross sections used in this benchmark problem are listed in Table 4. This problem was set up to analyze the features of spatial and angular discretization and demonstrate the parallel efficiency of the multithreaded strategy.
Region | |||
---|---|---|---|
R1 | 0.1 | 0.05 | 1.0 |
R2 | 0.0001 | 0.00005 | 0.0 |
R3 | 0.1 | 0.05 | 0.0 |
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F005.jpg)
Angular flux distribution over full domain with mesh sizes of 5 cm, 4 cm, 3 cm, 2.5 cm, and 2 cm are modeled and simulated. The quality of the quadrature sets was quantified by taking the root mean square (RMS) of the relative scalar flux of all N point detectors.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F006.jpg)
The ThorSNIPE code runs on a PC with 8 cores / 16 threads (one 11th Gen Intel(R) Core(TM) i7-11700F CPU running at 2.5 GHz) to demonstrate the performance and gain of the multithread technology. The speedup ratio with respect to serial mode versus the order of quadrature sets is presented in Fig. 6b, where the parallel implementation efficiency exceeds 100% when the S12 angular approximations are attained, and the mesh size is less than 3 cm. The comparative results show that the proposed method can achieve a better acceleration effect as the number of meshes increases. It can also be observed that the speedup ratio increases more slowly when the S20 angular approximations are attained. In this case, each thread has over 30 parallel tasks, and memory access becomes a bottleneck in the matrix solvers for partial differential equations. The numerical results are shown in Fig. 7, and the neutron fluxes at the key points are listed in Table 5. Compared with MCNP [31], the ThorSNIPE code effectively reduces angular discretization errors and fulfills the transport computational requirement to a certain degree.
Case | Co-ordinates | MCNP (cm-2s-1) | ThorSNIPE (cm-2s-1) | Relative errors |
---|---|---|---|---|
3A | (5,5,5) | 8.61578×10−0 | 8.61432×10−0 | −0.02 |
(5,15,5) | 2.16130×10−0 | 2.13602×10−0 | −1.17 | |
(5,25,5) | 8.93784×10−1 | 8.93551×10−1 | −0.03 | |
(5,35,5) | 4.78052×10−1 | 4.72878×10−1 | −1.08 | |
(5,45,5) | 2.89424×10−1 | 2.87933×10−1 | −0.52 | |
(5,55,5) | 1.92698×10−1 | 1.91669×10−1 | −0.53 | |
(5,65,5) | 1.04982×10−1 | 1.04268×10−1 | −0.68 | |
(5,75,5) | 3.37544×10−2 | 3.43534×10−2 | 1.77 | |
(5,85,5) | 1.08158×10−2 | 1.08608×10−2 | 0.42 | |
(5,95,5) | 3.39632×10−3 | 3.47318×10−3 | 2.26 | |
3B | (5,55,5) | 1.92698×10−1 | 1.91669×10−1 | −0.53 |
(15,55,5) | 6.72147×10−2 | 6.74001×10−2 | 0.28 | |
(25,55,5) | 2.21799×10−2 | 2.20850×10−2 | −0.43 | |
(35,55,5) | 9.90646×10−3 | 9.74980×10−3 | −1.58 | |
(45,55,5) | 3.39066×10−3 | 3.43623×10−3 | 1.34 | |
(55,55,5) | 1.05629×10−3 | 1.04113×10−3 | −1.44 | |
3C | (5,95,5) | 3.44804×10−4 | 3.45115×10−4 | 0.09 |
(15,95,5) | 2.91825×10−4 | 2.98723×10−4 | 2.36 | |
(25,95,5) | 2.05793×10−4 | 2.07437×10−4 | 0.80 | |
(35,95,5) | 2.62086×10−4 | 2.63330×10−4 | 0.47 | |
(45,95,5) | 1.05367×10−4 | 1.07620×10−4 | 2.14 | |
(55,95,5) | 4.44962×10−5 | 4.46921×10−5 | 0.44 |
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F007.jpg)
VENUS-3 benchmark
The VENUS-3 benchmark was proposed to validate the capabilities of the calculation methodologies and cross-section libraries for predicting fluence rates in RPVs because it has a very clean structural geometry representing standard PWR pressure vessel conditions [32]. Fig. 8 illustrates the geometric configuration and material distribution of the VENUS-3 facility. The VENUS-3 model includes two types of fuel rods: the PLSA, core baffle, water reflector, barrel, Pyrex control rods, barrel, neutron pad, three types of grids, bottom support, reactor vessel, and jacket inner walls. Detailed descriptions and qualifications were obtained from the benchmark document. The benchmark arranges 386 dosimeters, with 244 58Ni(n, p)58Co dosimeters, 104 115In(n, n’)115mIn dosimeters, and 38 27Al(n, a)24Na dosimeters placed in 268 different spatial locations. The experimental rate sets are provided as a ratio of the calculated reaction rate to the average dosimeter cross-section, named the equivalent fission flux.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F008.jpg)
The numerical results shown in Fig. 9 were modeled and simulated using the PNTN-S8 angular discretization, 199 KASHIL-E70 neutron groups multigroup cross-sections [33], and 407819 tetrahedral elements. Fig. 10 shows a C/E comparison of the equivalent fission fluxes at the three detector positions. For the 115In(n, n’)115mIn dosimeters, almost all the equivalent fission flux deviations are included within +/-10%, except for a detector position (No. 31 located in the core barrel with 0.7° angle) where the equivalent fission flux numerical value is affected by reflective boundary. For the 27Al(n, a)24Na dosimeters, 92% of the equivalent fission flux deviations were within +/-10%, and the remaining three dosimeters overestimated the corresponding experimental values by approximately 11.9%. For the 58Ni(n, p)58Co dosimeters, 95% of the equivalent fission flux deviations are limited within +/-10%, except for 11 detector positions where the equivalent fission flux numerical values overestimated the corresponding experimental values by about 10.1%-20.7%. They are concentrated at the junction of the PLSA region, outer baffle, and core barrel. The scalar flux distribution is not smooth because of the strong spatial nonuniformity of the geometry model, and higher requirements are proposed for local mesh refinement, which significantly degrades the simulated accuracy of these regions. The overall agreement between the numerical results from the ThorSNIPE code and the experimental results was very good, which shows that the ThorSNIPE code is highly reliable and stable for complex shielding calculations.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F009.jpg)
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F010.jpg)
The numerical results [32] from TORT-3.2 with 1004295 hexahedral grids and 26 BUGLE-96 neutron energy groups were selected as comparative references. For the uniformly distributed 3.3% fuel regions, TORT-3.2 with hexahedral structured meshes achieved better results. The ThorSNIPE code with tetrahedral unstructured meshes can decrease the numerical oscillation for complex regions. For the 27Al(n, a)24Na dosimeters, the numerical results of ThorSNIPE tended to be lower than those of TORT-3.2 in the core barrel. This underestimation might have been caused by an error in the average dosimeter cross section.
In Fig. 11, the speedup ratios of the three quadrature sets with the P3 Legendre expansion order and the four Legendre expansion orders with PNTN-S8 angular discretization are shown. The test conditions were consistent with TORT-3.2 code. Overall, the behavior of the proposed algorithm was similar to that of the Kobayashi-3i benchmark. For PNTN-S12, the acceleration ratio was 6.47 with 407819 elements. Comparing the results of different expansion orders, the P5 expansion order has the steepest upward trend, and the maximum speedup ratio achieved was 6.98. The numerical results indicate that the proposed algorithm achieves a favorable speedup ratio as the scale of the mesh increases, as expected.
202312/1001-8042-34-12-020/alternativeImage/1001-8042-34-12-020-F011.jpg)
Conclusion
In this study, a multithreaded parallel upwind algorithm to solve the first-order Boltzmann neutron equation is proposed and implemented in the multigroup 2D/3D SN transport code ThorSNIPE. The cell-based upwind sweep algorithm realizes angular flux transmission between neighboring meshes and achieves a stable solution by applying DGFEM spatial discretization. The multithread parallel strategy automatically manages the angular sweep stack of each thread and improves the quality and efficiency of the ThorSNIPE code.
The performance and accuracy of the proposed algorithm were tested using several typical benchmark problems. In the IAEA benchmark, the upwind sweep algorithm was proven to be the optimal method for DGFEM calculations and can save significant computational costs. The results show that the ThorSNIPE code agrees well with the reference, with a deviation of 1 pcm in eigenvalue and within 2.50% for flux distribution. In the Kobayashi-3i benchmark, the results for angular flux simulations over the full domain in different mesh sizes and quadrature set orders indicate that the multithreaded parallel algorithm has potential extensibility. In the VENUS-3 benchmark, 96% of the equivalent fission flux deviations are limited within +/-10%, except for 15 detector positions where the equivalent fission flux numerical values overestimated the corresponding experimental values by about 10.1%-20.7%. The maximum speedup ratio was 6.98 with 407819 elements. The results of the VENUS-3 benchmark show that this robust algorithm can play a reliable role in practical engineering applications and improve computational efficiency.
In the future, angular and algebraic multigrid methods planned to solve discretized equations to overcome the reduced convergence speed of the iteration method will be investigated. Furthermore, we plan to address more complex problems to assess the accuracy and effectiveness of the proposed method and present a rigorous mathematical convergence analysis of the new scheme.
New weighted-difference formulation for discrete-ordinates calculations. Tech. rep.
,ARES: a parallel discrete ordinates transport code for radiation shielding applications and reactor physics analysis
. Sci. Tech. Nucl. Install. 2017, 2596727 (2017). https://doi.org/10.1155/2017/2596727Reconstruction and parallelization of 3D SN program for neutron/photon transport
. Nucl. Power Eng. 35, 147-150 (2014). https://doi.org/10.13832/j.jnpe.2014.S2.0147A stabilized nodally integrated tetrahedral
. Int. J. Numer. Methods Eng. 67, 841-867 (2006). https://doi.org/10.1002/nme.1651Standard and goal-oriented adaptive mesh refinement applied to radiation transport on 2D unstructured triangular meshes
. J. Comput. Phys. 230, 763-788 (2011). https://doi.org/10.1016/j.jcp.2010.10.018Adaptive discontinuous finite element quadrature sets over an icosahedron for discrete ordinates method
. Nucl. Sci. Tech. 32, 98 (2021). https://doi.org/10.1007/s41365-021-00934-7VMultischeme equivalence procedure for neutron transport finite element methods
. Ann. Nucl. Energy 166, 108712 (2022). https://doi.org/10.1016/j.anucene.2021.108712A guide to numerical methods for transport equations
(Thesis,A new continuous finite element SN method for solving first-order neutron transport equation
. Ann. Nucl. Energy 175, 109237 (2022). https://doi.org/10.1016/j.anucene.2022.109237Streamline upwind Petrov–Galerkin methods for the steady-state Boltzmann transport equation
. Comput. Methods Appl. Mech. Eng. 195, 4448-4472 (2006). https://doi.org/10.1016/j.cma.2005.09.004Least-squares finite-element solution of the neutron transport equation in diffusive regimes
. SIAM J. Numer. Anal. 35, 806-835 (1998). https://doi.org/10.1137/S003614299629970Triangular mesh methods for the neutron transport equation. Tech. rep.
,Rattlesnake: A MOOSE-based multiphysics multischeme radiation transport application
. Nucl. Tech. 207, 1047-1072 (2021). https://doi.org/10.1080/00295450.2020.1843348Parallel algorithms for radiation transport on unstructured grids
. Proc. SC2000 Conf. (An algorithm for parallel SN sweeps on unstructured meshes, Nucl
. Sci. Eng. 140, 11-136 (2002).Minaret, a deterministic neutron transport solver for nuclear core calculations
. International Conference on Mathematics and Computational Methods Applied to Nuclear Science and Engineering (A deep penetration problem calculation using AETIUS: an easy modeling discrete ordinates transport code using unstructured tetrahedral mesh, shared memory parallel
. EPJ Web of Conferences 153, 06025 (2017). https://doi.org/10.1051/epjconf/201715306025A two-dimensional finite element shielding calculation code with mass-matrix lumping technique and unstructured meshes
. Nucl. Tech. 46, 77-85 (2023). https://doi.org/10.11889/j.0253-3219.2023.hjs.46.020602. (in Chinese)The deal. II library, version 9.4
. J. Numer. Math. 30, 231-246 (2022). https://doi.org/10.1515/jnma-2022-0054Globally conservative, hybrid self-adjoint angular flux and least-squares method compatible with voids
. Nucl. Sci. Eng. 185, 294-306 (2017). https://doi.org/10.1080/00295639.2016.1272374Performance and performance modeling of a parallel algorithm for solving the neutron transport equation
. J. Supercomp. 6, 211-235 (1992). https://doi.org/10.1007/BF00155800Neutronic calculations of the China dual-functional lithium–lead test blanket module with the parallel discrete ordinates code Hydra
. Nucl. Sci. Tech. 31, 74 (2020). https://doi.org/10.1007/s41365-020-00789-4A parallel algorithm with interface prediction and correction for spherical geometric transport equation
. Prog. Nucl. Energy 51, 268-273 (2009). https://doi.org/10.1016/j.pnucene.2008.09.003Comparison via parallel performance models of angular and spatial domain decompositions for solving neutral particle transport problems
. Prog. Nucl. Energy 47, 37-60 (2007). https://doi.org/10.1016/j.pnucene.2006.08.003On the adequacy of message-passing parallel supercomputers for solving neutron transport problems
. In Proceedings of the 1991 ACM/IEEE Conference on Supercomputing (Calculation of four thermal reactor benchmark problems in X-Y geometry. Tech. Rep
.3D radiation transport benchmark problems and results for simple geometries with void region
. Prog. Nucl. Energy 39, 229-144 (2001). https://doi.org/10.1016/S0149-1970(01)00007-5ENEA nuclear data center neutron transport analysis of the VENUS-3 shielding benchmark experiment. Tech. Rep
.Validation of an ENDF/B-VII.0-based neutron and photon shielding library in MATXS-format
. J. Korean Phy. Soc. 59, 1199-1202 (2011). https://doi.org/10.3938/jkps.59.1199The authors declare that they have no competing interests.