Introduction
Nuclear science and technology are rapidly developing and have been applied in various fields, having increasingly important roles in the sphere of scientific research and production[1-3]. Simultaneously, menacing nuclear weapons and radioactive contamination by nuclear industrial accidents present long-term and significant consequences for the environment, ecology, and biological health[4-6]. The detection and identification of radionuclides are crucial tasks under such circumstances [1, 7]. It is important to develop effective algorithms for the detection and identification of radionuclides with stronger discrimination and high accuracy. One of the most critical steps in radionuclide identification is the feature extraction from gamma-ray energy spectra, which is a complicated task owing to the background conditions, energy resolution of the radiation detector, calibration shift, characteristic peak overlap, source strength, and shielding status[8, 9].
Feature extraction methods of traditional radionuclide identification algorithms can be summarized as searching for characteristic energy peaks from the gamma-ray energy spectra and matching them with peaks in the radionuclide library[2, 4, 8-11], which are usually based on physical rules and do not require a training process. Another classical approach is the template matching method, for which the main idea is establishing a template library of the gamma-ray energy spectra in advance and matching the entire spectrum or transformation of the spectrum with the template in the library[12, 13]. These methods are highly operator-dependent, and their limitations are magnified when characteristic peaks are overwhelmed by background noise or several interfering peaks are extracted.
With the development of artificial intelligence, radionuclide identification has gradually become a widely studied classification problem[14-16]. The main idea of classification methods applied in radionuclide identification is to extract features from gamma-ray energy spectra of known types, and train classification models using extracted features; then, the trained model is applied to estimate the probability of existing radionuclides of unknown types[17]. Numerous feature extraction algorithms have been used for radionuclide identification, such as the Karhunen-Loeve transform (K-L transform)[18], principal component analysis (PCA)[13, 19], singular value decomposition (SVD)[20], wavelet [21, 22], discrete cosine transform (DCT)[23], and sparse representation [24]. Subsequently, classification methods such as Bayesian[23, 25], extreme gradient boosting tree [26], back propagation neural network [17], artificial neural network (ANN)[27, 28], fuzzy logic [29], long short-term memory (LTSM)[30], convolutional neural network (CNN)[31], and deep convolutional neural network (DCNN)[32, 33] were applied for radionuclide identification. The key factor for the success of these methods is the extraction of strong discriminative features. The limitation of the aforementioned features is that only the relative relationship between the corresponding counts of the front and rear energy addresses is considered, and errors may increase owing to the non-smoothness of the low-count spectra[34-36].
Recent studies have shown that image descriptors transferred by CNNs and DCNNs provide a stable and reliable performance for image classification problems[37-44]. Hu et al. reported that features transferred from CNNs were sufficiently generalized to high-resolution remote sensing image datasets and were more expressive than low-level and mid-level features[37]. Babenko et al. experimentally determined that the activation of the top layers of CNNs is competitive despite being trained for unrelated classification tasks such as ImageNet[38]. Moreover, Y. Gong et al. transferred the outputs of the last fully connected layer of a DCNN as an image descriptor[39], and Razavian et al. demonstrated that features transferred from convolution layers can provide useful global descriptors of specific image regions[40-44]. Liu et al. demonstrated that convolution layers have excellent generalization and efficiency and that transferring convolution layer features can achieve an advanced performance[44].
This study constructed a novel feature extraction method from gamma-ray energy spectra for radionuclide identification. First, the gamma-ray energy spectra are transformed from a vector to matrix and then to image form. Feature transferring is then performed using a DCNN model. The transferred image descriptors consist of the activations from the convolution layers and activation vectors of the fully connected layers. To verify the effectiveness of the proposed method, 11 classical classification methods were employed to perform a statistical comparison, and the results demonstrate that the proposed method significantly outperforms the peak-searching-based method, histogram of oriented gradients (HOG), and scale-invariant feature transform (SIFT).
The following two main contributions are presented in this study:
• A novel pre-process method of the gamma-ray energy spectra is proposed. The vectors of the gamma-ray energy spectra are mapped to matrix form, and further mapped to image form. This form conversion can improve the utilization of spectral information and serve as the basis for extracting essential features and constructing a more discriminative classifier.
• Exploring and verifying the application of image descriptors transferring from a DCNN model in the field of radionuclide identification. Experimental results demonstrate that image descriptors can effectively extract the essential features of the gamma-ray energy spectrum images. Local image descriptors transferred from higher convolution layers can provide more discriminative descriptors, and global image descriptors transferred from the first fully-connected layer has the strongest semantic information among the fully-connected layer.
The remainder of this study is organized as follows. In Sect. 2, we introduce the proposed feature-extraction approach for radionuclide identification. Section 2 presents a series of experiments using both synthetic datasets and measured datasets from real laboratory environments and offers a comparative analysis. Section 3 concludes the study.
Method
The proposed method consists of the following two major steps: (a) Mapping the vectors of the gamma-ray energy spectra from a Euclidean space to matrices in a Banach space, and then mapping the matrices to images in the RGB color space. (b) Constructing a DCNN model trained on ImageNet and transferring the corresponding activation vectors of fully connected layers and activations of convolution layers as global and local features of the gamma-ray energy spectrum images. Fig. 1 presents the procedure of the proposed method.
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F001.jpg)
Data mapping
In this subsection, the essential features of the different radionuclides are extracted from a novel perspective.
Gamma rays are the products of the de-excitation process of atomic excitation and manifest as short-wavelength electromagnetic radiation. The essence of gamma rays is a stream of particles, which are gamma photons. Gamma photons are uncharged particles, and their interaction with matter is a random event. Collecting gamma photons using a general signal acquisition device is complex, whereas collecting electrical signals is relatively easy; the amplitude of the electrical signal is proportional to the photon energy value. Gamma photons can be converted into electrical signals for signal processing to be collected by a nuclear radiation detector. The principle of this process is that photons emitted by the radioactive source interact with the atoms of the medium in the detector to produce charged particles. The detector collects the particles, converts them into electrical signals, and detects nuclear signals by measuring the electrical signals. The counts distributed with the energy value of the particles can be obtained by scaling the pulse amplitude by the energy; the energy spectrum is the curve of the distribution of the counts with the energy of the particles. As the fingerprints of radionuclides, the energy spectrum contains the distinguishable information of different radionuclides[2-4, 6].
For nuclear events, the count and time of events are random within a certain time interval. In radiation detection, the number of nuclear events measured over a certain period (e.g., detector counts) is also random. Because radioactive decay is a random process, each observation can be considered as a random experiment, and the count per unit time can be regarded as a random variable that obeys the Poisson or Gaussian distributions. For a spherical space with a single point source as the center of the sphere and a certain radius, the process of generating gamma photons by radioactive decay is random and continuous, and photons are uniformly emitted in all directions in space.
The gamma-ray energy spectrum 𝒔 is a broad stationary random vector in Euclidean space,
Considering 𝒔 as a Markov chain or Markov process, which is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event, sk is only related to sk-1. In vector-based feature extraction methods, only part of the information of 𝒔 is considered, that is, the relative relationship between the corresponding counts of the front and rear energy addresses. This is not conducive to mining the mutual relationship between non-adjacent sk, which will cause difficulties in extracting discriminative and effective features.
Every radioactive decay produces photons of different energy values, and the photon counts of a certain energy obey the Poisson or Gaussian distributions. Thus, the total count of photons generated by radioactive decay per unit time also obeys the Poisson or Gaussian distributions. Let 𝒂 and 𝒃 be the gamma-ray energy spectra obtained by using different measurement durations in a measurement scenario with a fixed background noise, using a specified detector to measure the specified radioactive source at a fixed location and orientation. Therefore, as previously indicated, 𝒂 and 𝒃 are vectors from the same 𝑯,
To efficiently transfer discriminative information for identification from the gamma-ray energy spectra, mapping the gamma-ray energy spectra from a vector form in 𝑯 to a matrix form in Banach space 𝑩, the mapping f can be formulated by Eq. (4).
From the matrix perspective, there are more elements adjacent to pi j. When extracting features from gamma-ray energy spectra, not only is the relative relationship between the corresponding counts of the front and rear energy addresses considered, but also the relative relationship between the upper and lower counts and the diagonal counts, which can be easily used and is more conducive to mining the internal and mutual relationships between elements in 𝑷.
To apply 𝑷 as the input of a DCNN model and transfer image descriptors as features of the gamma-ray energy spectra, it is essential to map 𝑷 to the image form. Mapping g maps 𝑷 from the matrix form in 𝑩 to the image form in the RGB color space 𝑱, can be formulated by Eq. (6).
Eq. (6) maps the element values of 𝑸 onto the corresponding pixels of an image with specified colors. Each qi j corresponds to a rectangular area in the image, and the values of qi j are indices in the Parula color map [45] that determine the color of each patch. Eq. (6) maps the smallest value in 𝑸 to the first entry in the Parula color map and maps the largest value in 𝑸 to the last entry in the Parula color map. All intermediate values of 𝑸 are linearly scaled to the Parula color map in the ascending order. The relationship between the values of the elements in 𝑸 and the colors of the corresponding pixels in the Parula color map is shown in Fig. 2.
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F002.jpg)
𝒂 and 𝒃 are two gamma-ray energy spectra obtained for different measurement durations under the same measurement conditions, and Qa=g(f(a)) and Qb=g(f(b)) are the images of 𝒂 and 𝒃, respectively, under mapping f * g which are defined as previously mentioned. For
Based on the aforementioned analysis, the effects of different measurement durations on 𝑸a and 𝑸b can be ignored under the same measurement conditions. More explicitly, the mappings of the gamma-ray energy spectra in the RGB color space from the same radionuclide would present nearly identical images in an ideal situation, which can reduce the intraclass differences caused by different measuring durations. Even in a real measuring situation, the information processing of mapping the gamma-ray energy spectra from a Euclidean space to a Banach space and then to an RGB color space remains to have a strong applicability and can reduce intraclass differences, thereby helping to extract the essential features from the gamma-ray energy spectra of different radionuclides.
Feature transferring
In this subsection, a DCNN model was constructed and trained on the ImageNet dataset. The mapped gamma-ray energy spectrum images were applied as inputs to the DCNN model, and the corresponding activation vectors of fully connected layers and activations from convolution layers were transferred as descriptors of images to construct a new classification model for radionuclide identification.
VGG is a widely used convolutional neural network (CNN) model proposed by Karen Simonyan and Andrew Zisserman at the University of Oxford[46]. The VGG has various configurations; instance, VGG-11, VGG-16, VGG-19, etc.. Of all the configurations, VGG-16 was identified as the best-performing model on the ImageNet dataset. The basic building block of VGG can be summarized as a stack of multiple (usually one, two, or three) convolution layers with a filter size of 3 × 3, one stride, and one padding, followed by a max-pooling layer of size 2 × 2. Different configurations of this stack were repeated in the network to achieve various depths. The number associated with each configuration is the number of layers with the weight parameters. The convolution stacks are followed by three fully connected layers, two with a size of 4096, and the last one with a size of 1000. The last layer is the output layer with a Softmax activation. A size of 1000 refers to the total number of possible classes in ImageNet. In the method proposed in this study, the structure of the VGG consists of U convolution layers, V max pooling layers, and N fully connected layers. The convolution layers use convolution kernels to convolve the input image, and the results of the convolution constitute the feature maps of the input image; thus, the local features of the image are extracted after the convolution layers. The max pooling layers are arranged after the convolution layers, and the maximum values of the corresponding positions in the feature maps are calculated. This can reduce the dimension of the extracted feature information to make the feature maps smaller, simplify the computing complexity of the network, and avoid overfitting. The U convolution layers can be divided into M groups using the max pooling layers as separations. The fully connected layers map the distributed feature representation learned from the convolution layers and max pooling layers to the sample marker space, the essence of which is to perform a weighted sum of the features to integrate the local features and output them as a value to reduce the influence of the local feature position on the classification.
The convolution layers and max pooling layers have parameters such as the height of the convolution kernel, width of the convolution kernel, number of input channels, number of output channels (number of convolution kernels), padding, and stride. The padding parameter refers to the boundary padding of the original matrix with ′0′ before the convolution; thus, the convolution kernel can extend to the pseudo-pixels beyond the edge when scanning the input image, thereby avoiding the loss of edge information. The stride parameter refers to the length of each movement of the convolution kernel, and the stride size affects the efficiency of the model. Each convolution kernel strictly has a bias parameter. To simplify the calculation, the bias is omitted in the following calculation.
For the convolution layers, the height and width of the convolution kernel are represented as
The size of the input image is represented as
VGG is training on the ImageNet dataset to obtain the classification model PM. ImageNet is an image database organized according to the WordNet hierarchy, in which hundreds or thousands of images depict each node of the hierarchy. The dataset has been instrumental in advancing computer vision and deep learning research[47]. Let 𝑺 be a set of the gamma-ray energy spectra in Euclidean space,
The convolution kernel of PM corresponds to a receptive field, and a small part of the image (local receptive area) is used as the input of the lowest convolution layer, which makes each neuron output by the convolution layer only experience the local image area and does not need to experience the global image. This operation is equivalent to passing through a digital filter to obtain the most salient features of the observed data. In the fully connected layers, different local features from the convolution layers are synthesized through the weight matrix to form a representation of the global information. Therefore, the transferred activation vectors of the fully connected layers are regarded as global feature representations for the gamma-ray energy spectrum images, and the transferred activation maps from the convolution layers are regarded as local features describing particular gamma-ray energy spectrum image regions.
Specifically, the activated maps transferred from the convolution layers, that is, the set of feature maps transferred from the i-th group of convolution layers of PM, is denoted as 𝒄i,
The transferred activation vectors of the fully connected layers, that is, the feature vectors transferred from the j-th fully connected layer of
Algorithm: | Feature Extraction Approach based on Image |
---|---|
Descriptor Transferring | |
Input: | |
s: Gamma-ray energy spectrum; | |
n: Number of columns of 𝑷 and 𝑸; | |
cm: Parula colormap, a 64× 3 matrix; | |
VGG: A DCNN framework with M groups of | |
convolution Layers and N fully-connected layers; | |
ImageNet: Public image data set; | |
Begin: | |
1. 𝒔 is mapped to matrix form by: | |
where i=0,1, ..., m-1, j=0,1, ..., n-1; | |
2. pi j is normalized by: | |
where i=0,1, ..., m-1, j=0,1, ..., n-1; | |
3.qi j is mapped to image form by: | |
where |
|
ls means linearly scaled; | |
4. |
|
5. Pre-trained model is obtained based on Imagenet: | |
PM = VGG (ImageNet) | |
6. 𝒄i is output of the i-th group of convolution layers of PM, where |
|
7. Converting the set of feature maps into set of feature vectors by |
|
8. 𝒇j is output of the i-th fully-connected layer of PM, where |
|
End | |
Output: | |
Transferred local features of 𝒔: | |
Transferred global features of 𝒔: | |
Illustration
In this subsection, we present a simple example of the proposed method. For illustration, we selected two synthetic gamma-ray energy spectra 𝒔1 and 𝒔2 of 60Co. The essence of 𝒔1 and 𝒔2 are two vectors in Euclidean space 𝑯l, where l = 4096. The simulation settings of 𝒔1 and 𝒔2 are essentially the same, and the only difference is the total number of simulated particles, which is equivalent to the difference in the measurement duration. The number of simulated particles correlates with the duration of the measurement because the process of generating gamma photons by radioactive decay is random and continuous, and photons are uniformly emitted in all directions in space. The simulated particle numbers corresponding to 𝒔1 and 𝒔2 are represented as t1 and t2, where t1 is five hundred thousand and t2 is two hundred and fifty thousand. To illustrate the difference between the two spectra more vividly, we plot 𝒔1 and 𝒔2 in Fig. 3, where the blue spectrum is the description of 𝒔1, and the red spectrum is the description of 𝒔2. Figure. 3 clearly shows differences between 𝒔1 and 𝒔2 due to the duration of measurement, where the counts distributed over the energy addresses of the entire spectrum is significantly more prominent in 𝒔1 than 𝒔2.
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F003.jpg)
To extract more discriminative image descriptors from the gamma-ray energy spectra for identification, 𝒔1 and 𝒔2 are mapped to the matrix form using Eq. (4). The mapped matrices are represented as 𝑷1 and 𝑷2, which are in the Banach space
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F004.jpg)
In the feature transfer phase, the structure of VGG and the corresponding feature sizes are listed in Table 2, which consists of five groups of convolution layers and two fully connected layers. The layer number only considers the number of convolution layers, and the last layer of each group is the max-pooling layer. The number of channels in the feature size is not changed in the convolution layers, but changes in the max pooling layer owing to the size of the stride parameter (the number of output channels in different groups can be calculated by Eq. (9), and Eq. (10)). For fully connected layers, the number of output channels is equal to the number of neurons in the fully connected layers. The partial parameters of the convolution, max pooling, and fully connected layers are shown in Table 3.
Layer group | Layer number | Feature size |
---|---|---|
Conv1 | 2 | 64×64×64 |
Conv2 | 2 | 128×32×32 |
Conv3 | 3 | 256×16×16 |
Conv4 | 3 | 512×8×8 |
Conv5 | 3 | 512×4×4 |
FC | 2 | 4096 |
Output | 1 | 1000 |
Layer name | Kernel size/Neurons number | Stride | Padding |
---|---|---|---|
Convolution | 3 × 3 | 1 | 1 |
Max pooling | 2 × 2 | 2 | 0 |
Fully-connected | 4096 | - | - |
The VGG framework was trained on the ImageNet dataset to obtain the classification model PM. 𝑸1 and 𝑸2 are applied as input images for the PM to transfer features, and the activation vectors of the fully connected layers and activations from the convolution layers in the process are transferred as the image descriptors of 𝑸a and 𝑸b. Image descriptors transferred from the fifth group of convolution layers 𝒄5 and the first fully connected layer 𝒇1 were selected as illustrations. 𝒄5 and 𝒇1 from the synthetic and measured datasets are shown in a low-dimensional space through the t-SNE[48] in Fig. 7. Each color in the figure represents one type of radionuclide, which intuitively reflects the difference in the expressive and discriminative abilities between the different features. The transferred image descriptors have a strong intra-class similarity and inter-class differentiation, and a further analysis of the feature performance is presented in the next section.
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F007.jpg)
Experiments and Analysis
This section introduces the acquisition and preprocessing of the synthetic data and the measured data and establishes a series of experiments using 28 classification methods based on the Weka machine learning toolkit to verify the feasibility of image descriptors transferred from the gamma-ray energy spectrum for radionuclide identification. Based on the previous experiments, statistical comparisons of features through non-parametric and Friedman tests were conducted to verify whether image descriptors transferred from DCNNs can be used as an essential feature representation for gamma-ray energy spectrum images.
Data preparation
The production of the synthetic dataset consisted of the following three steps: (1) Background data acquisition. A self-made 3-inch NaI detector was used to measure the ambient background. The measurement was conducted in a laboratory environment without the presence of a separate radioactive source. The detector was then placed at a fixed position for 12 h. Two measurements were performed, one with lead bricks placed around the detector and the other without; two sets of background data were obtained. (2) Single-nuclide energy-spectrum acquisition. Based on the Geant4 platform, the transport process of the radioactive gamma-ray particles of 26 single nuclides was simulated using the Monte Carlo method. We constructed simulated scenarios in Geant4 containing only different types of single radioactive point sources and a 3-inch NaI detector. The positions and relative distances of the radioactive point sources and detector were fixed. A total of one million photons were simulated, and their trajectories were recorded by the detector and converted into the gamma-ray energy spectrum. (3) Data synthesis. To simulate real measurements as closely as possible, two sets of background data and 26 sets of synthetic spectra were linearly superimposed with a random signal-to-noise ratio (SNR).
Type | Radionuclide |
---|---|
SNM | 237NP, 233U, 235U, 238U |
Industrial | 241Am, 133Ba, 57Co, 60Co, 137Cs, |
152Eu, 192Ir, 75Se | |
Medical | 51Cr, 18F, 67Ga, 123I, 125I, 131I, |
111In, 103Pd, 99mTc, 201Tl, 133Xe | |
NORM | 40K, 226Ra, 232Th |
The measured gamma-ray energy spectra were obtained from radioactive sources in a laboratory environment with lead brick shielding. The measurements took advantage of a cadmium zinc telluride (CZT) gamma-ray spectrometer from Kromek. The spectrometer has 4096 measuring channels; the measurable energy ranges between 25 keV-3.0 MeV, and the electronic noise is lower than 10 keV. Two types of V radiation sources, 137Cs and 60Co, and one type of IV radiation source, 152Eu were used in the measurement. The spectrometer was carried on the Turtlebot robot to quantitatively control the measuring distance and reduce radiation damage to the experimental operators. As shown in Table 5, seven groups of samples were established; the measuring object, measuring distance, and measuring duration were varied in the process. A total of 150 measured gamma-ray energy spectra of three single radiation sources and a total of 200 measured gamma-ray energy spectra of four mixed radiation sources were obtained and named dataset 2. Figure 5 presents the spectrum of 60Co in different sample sets. Figure 6 presents the gamma-ray energy spectrum images in different sample sets.
Set name | Nuclide type | Capacity |
---|---|---|
Data set 1 | Radionuclides listed in Table 4 | 2080 |
Data set 2 | 60Co | 50 |
137Cs | 50 | |
152Eu | 50 | |
137Cs, 60Co | 50 | |
137Cs, 152Eu | 50 | |
60Co, 152Eu | 50 | |
137Cs, 60Co,152Eu | 50 |
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F005.jpg)
-202212/1001-8042-33-12-008/alternativeImage/1001-8042-33-12-008-F006.jpg)
Feature performance comparison and analysis
Research have shown that image descriptors transferred from DCNNs can provide a reliable performance for image classification problems. However, choosing different features for a specific classification domain remains worth discussing. For the domain of the radioactive gamma-ray energy spectral classification presented in this study, and considering the limitation of the computing resources and the scale of parameters, VGG-16 was chosen as the DCNN framework owing to the characteristic that VGG-16 applies a significantly small 3 × 3 receptive field (filters) throughout the entire network with the stride of 1 pixel. A combination of multiple 3 × 3 filters and nonlinear activation layers can replace a receptive area of a larger size, which makes the decision functions more discriminative to the characteristics of the spectrum. This imparts the ability of the network to converge faster[46]. In addition, the consistent use of 3 × 3 convolutions across the network makes the network significantly simple, elegant, and conveniently transfers image descriptors.
In this subsection, 28 classification methods based on the Weka machine-learning toolkit [49] were applied in a series of experiments to verify whether image descriptors transferred from a DCNN model can be used to construct a classification model with a strong discrimination and advanced accuracy. Owing to the large number of transferred features (five sets of local features and two sets of global features for a single energy spectrum), in an actual training process, using only a certain set of features is sufficient to build a satisfactory classifier. Therefore, in the experiments described in this subsection, single sets of local and global features and combinations of high-level features were used to conduct the experiments. Multiple groups of transferred image descriptors from the synthetic and measured datasets were applied in the training and testing processes of the 28 classification models. All classification models were trained with the default parameters and settings specified in the toolkit, and the experiments applied the ten-fold cross-validation method to avoid the imbalance caused by random data segmentation. The percentage of misclassified cases for transferred features on the synthetic dataset and measured dataset are listed in Table 6 and Table 7. The results demonstrate that although the gamma-ray energy spectrum images are completely unfamiliar to the DCNN model and have not been used in the pre-training process, the transferred image descriptors achieved good classification results.
Classification Method | 𝒄1 | c2 | c3 | c4 | c5 | c4,5 | c5,4 | c3,4,5 | c5,4,3 | 𝒇1 | 𝒇2 | Q1,2 |
---|---|---|---|---|---|---|---|---|---|---|---|---|
Bayes Net | 16.68 | 5.62 | 5.62 | 1.20 | 45.14 | 0.43 | 0.43 | 0.87 | 0.82 | 0.43 | 68.80 | 0.48 |
Lib SVM | 56.87 | 39.04 | 39.04 | 32.36 | 27.40 | 29.47 | 29.57 | 30.43 | 30.48 | 15.38 | 97.12 | 34.62 |
RBF Classifier | 70.38 | 69.37 | 69.37 | 75.34 | 71.39 | 72.07 | 72.12 | 72.98 | 72.60 | 69.23 | 80.12 | - |
RBF Network | 11.01 | 3.03 | 3.03 | 0.38 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 85.34 | 32.45 |
Simple Logistic | 2.07 | 0.05 | 0.05 | 0.10 | 0.00 | 0.24 | 0.24 | 0.00 | 0.24 | 0.00 | 69.95 | 0.00 |
SMO | 15.53 | 0.05 | 0.05 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 57.31 | 0.24 |
IBk | 1.73 | 0.05 | 0.05 | 0.00 | 0.10 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 81.35 | 1.44 |
LWL | 38.37 | 24.71 | 24.71 | 15.72 | 14.76 | 33.41 | 13.65 | 14.47 | 14.47 | 8.37 | 85.58 | 34.86 |
Attribute Selected Classifier | 9.62 | 1.54 | 1.54 | 1.25 | 1.15 | 0.91 | 0.96 | 0.96 | 0.72 | 1.43 | 85.10 | 1.43 |
Bagging | 4.37 | 1.06 | 1.06 | 0.29 | 3.70 | 0.38 | 0.19 | 0.38 | 0.14 | 0.24 | 82.69 | 0.48 |
Classification Via Regression | 3.27 | 1.59 | 1.59 | 0.53 | 1.15 | 0.38 | 0.82 | 4.00 | 0.82 | 0.96 | 82.69 | 2.88 |
Filtered Classifier | 8.89 | 6.59 | 6.59 | 0.53 | 6.54 | 6.39 | 7.26 | 6.01 | 5.72 | 6.63 | 86.92 | 7.74 |
Iterative Classifier Optimizer | 2.98 | 0.72 | 0.38 | 0.24 | 0.00 | 0.00 | 34.86 | 0.00 | 0.00 | 0.72 | 78.85 | 0.72 |
Logit Boost | 2.98 | 0.38 | 0.38 | 0.24 | 0.19 | 0.14 | 0.00 | 0.00 | 0.00 | 0.48 | 80.71 | 0.00 |
MultiClass Classifier | 2.31 | 0.05 | 0.05 | 0.10 | 0.24 | 0.00 | 0.00 | 0.00 | 0.00 | 0.24 | 86.54 | 0.00 |
MultiClass Classifier Updateable | 0.00 | 0.34 | 0.34 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.24 | 82.21 | 3.12 |
Random Committee | 2.50 | 0.24 | 0.24 | 0.05 | 0.19 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 87.02 | 0.00 |
Randomizable Filtered Classifier | 4.86 | 0.34 | 0.34 | 0.24 | 0.10 | 0.10 | 0.19 | 0.14 | 0.19 | 0.24 | 93.51 | 90.19 |
Random Sub Space | 3.51 | 0.77 | 0.77 | 0.05 | 0.19 | 0.10 | 0.05 | 0.10 | 0.19 | 0.24 | 83.89 | 0.19 |
Decision Table | 20.38 | 17.93 | 17.93 | 12.93 | 0.34 | 12.31 | 12.36 | 15.91 | 18.27 | 21.87 | 88.70 | 21.87 |
JRip | 8.46 | 5.34 | 5.34 | 3.70 | 3.75 | 12.31 | 4.57 | 0.00 | 6.01 | 0.19 | 88.46 | 0.00 |
OneR | 41.44 | 41.39 | 41.39 | 45.14 | 48.12 | 43.17 | 43.17 | 41.54 | 42.12 | 64.66 | 94.23 | 64.71 |
PART | 4.95 | 1.39 | 1.39 | 1.15 | 1.59 | 1.30 | 1.15 | 1.25 | 1.06 | 2.36 | 88.46 | 1.59 |
J48 | 4.86 | 1.30 | 1.30 | 0.62 | 1.25 | 0.62 | 0.62 | 0.87 | 0.53 | 1.20 | 91.11 | 1.44 |
LMT | 0.00 | 0.00 | 0.05 | 0.00 | 0.48 | 0.48 | 0.48 | 0.24 | 0.24 | 0.00 | 68.92 | 0.00 |
Random Forest | 2.07 | 0.24 | 0.24 | 1.73 | 0.24 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 79.52 | 0.05 |
Random Tree | 6.73 | 2.93 | 2.93 | 1.73 | 3.12 | 2.84 | 2.26 | 2.21 | 2.45 | 4.66 | 92.55 | 5.19 |
REP Tree | 6.39 | 2.79 | 2.79 | 2.16 | 2.16 | 1.87 | 1.35 | 1.49 | 1.44 | 3.08 | 86.30 | 3.08 |
Mean | 12.61 | 8.17 | 8.16 | 7.06 | 8.33 | 7.82 | 8.08 | 6.92 | 7.09 | 7.24 | 83.36 | 11.44 |
Classification method | c1 | c2 | c3 | c4 | c5 | c4,5 | c5,4 | c3,4,5 | c5,4,3 | 𝒇1 | 𝒇2 | Q1,2 |
---|---|---|---|---|---|---|---|---|---|---|---|---|
Bayes net | 6.57 | 5.71 | 6.00 | 8.29 | 8.29 | 8.57 | 8.57 | 8.00 | 8.00 | 4.86 | 32.86 | 32.86 |
Lib SVM | 54.29 | 54.29 | 52.57 | 45.14 | 40.00 | 41.43 | 41.43 | 41.14 | 41.14 | 30.00 | 36.00 | 30.57 |
RBF classifier | 38.29 | 45.71 | 33.71 | 30.29 | 26.00 | 36.29 | 27.14 | 33.43 | 34.29 | 22.86 | 61.43 | - |
RBF network | 18.86 | 22.57 | 9.71 | 6.29 | 2.86 | 3.43 | 3.43 | 3.71 | 3.71 | 1.43 | 65.71 | 33.14 |
Simple logistic | 9.71 | 4.86 | 2.57 | 2.29 | 3.14 | 1.14 | 3.43 | 1.14 | 1.14 | 0.57 | 34.29 | 0.57 |
SMO | 24.00 | 14.57 | 1.43 | 0.86 | 0.57 | 0.57 | 0.57 | 0.86 | 0.86 | 0.57 | 22.57 | 22.57 |
IBk | 2.57 | 2.00 | 1.14 | 0.86 | 0.86 | 0.57 | 0.57 | 0.86 | 0.57 | 0.00 | 42.86 | 42.86 |
LWL | 39.43 | 30.00 | 35.43 | 18.86 | 16.00 | 18.29 | 18.29 | 18.29 | 18.29 | 11.43 | 62.86 | 63.71 |
Attribute selected classifier | 5.14 | 7.43 | 5.71 | 6.57 | 4.29 | 6.86 | 6.57 | 5.43 | 6.00 | 6.00 | 65.71 | 1.43 |
Bagging | 3.71 | 2.86 | 4.86 | 3.14 | 2.86 | 3.43 | 2.00 | 4.57 | 2.29 | 2.00 | 47.14 | 34.29 |
Classification via regression | 4.00 | 4.00 | 3.71 | 4.00 | 4.86 | 4.57 | 4.29 | 4.00 | 4.57 | 2.00 | 52.74 | 2.00 |
Filtered classifier | 5.14 | 5.71 | 5.43 | 6.29 | 9.43 | 5.71 | 6.00 | 6.00 | 6.00 | 62.57 | 62.57 | 62.86 |
Iterative classifier optimizer | 3.43 | 2.86 | 2.86 | 2.86 | 2.57 | 2.00 | 4.00 | 2.57 | 2.57 | 1.71 | 51.14 | 0.00 |
Logit boost | 3.43 | 2.57 | 3.43 | 2.86 | 2.29 | 2.00 | 2.00 | 2.57 | 2.57 | 0.57 | 42.86 | 0.29 |
MultiClass classifier | 9.71 | 5.43 | 3.14 | 2.29 | 1.43 | 1.43 | 1.14 | 1.71 | 2.00 | 2.00 | 54.28 | 8.57 |
MultiClass classifier updateable | 38.29 | 14.29 | 4.00 | 4.00 | 1.43 | 1.71 | 1.71 | 2.29 | 1.71 | 4.00 | 35.71 | 4.29 |
Random committee | 2.00 | 2.57 | 1.71 | 2.29 | 2.00 | 2.00 | 1.14 | 1.14 | 1.14 | 1.14 | 55.71 | 0.57 |
Randomizable filtered classifier | 4.57 | 5.14 | 3.71 | 3.71 | 2.57 | 2.57 | 3.14 | 3.14 | 3.14 | 2.00 | 79.14 | 61.14 |
Random sub Space | 3.43 | 2.00 | 1.71 | 2.00 | 2.00 | 2.00 | 4.00 | 3.14 | 3.14 | 1.71 | 48.86 | 79.71 |
Decision table | 5.43 | 8.29 | 8.86 | 9.43 | 6.57 | 9.71 | 8.00 | 10.29 | 9.14 | 8.00 | 66.86 | 8.00 |
JRip | 6.86 | 6.57 | 5.14 | 5.71 | 9.43 | 4.29 | 6.29 | 6.86 | 6.57 | 6.00 | 63.43 | 4.29 |
OneR | 8.86 | 11.72 | 9.14 | 8.57 | 12.57 | 9.71 | 9.71 | 9.71 | 9.71 | 14.29 | 79.71 | 14.29 |
PART | 3.71 | 5.72 | 5.43 | 5.14 | 4.57 | 4.00 | 4.29 | 6.00 | 5.14 | 2.00 | 63.71 | 2.00 |
J48 | 3.14 | 4.86 | 4.86 | 4.00 | 4.29 | 3.43 | 3.43 | 4.86 | 4.57 | 2.57 | 72.00 | 2.57 |
LMT | 2.57 | 4.57 | 2.57 | 2.29 | 0.57 | 1.14 | 0.86 | 1.14 | 1.14 | 0.57 | 32.86 | 0.57 |
Random forest | 2.29 | 2.29 | 1.14 | 0.86 | 1.14 | 0.29 | 0.57 | 0.57 | 0.57 | 0.57 | 40.57 | 0.57 |
Random tree | 4.29 | 4.57 | 4.57 | 5.71 | 4.00 | 4.57 | 5.14 | 4.57 | 3.71 | 2.86 | 74.57 | 4.00 |
REP tree | 5.14 | 6.86 | 7.71 | 5.43 | 5.14 | 4.86 | 5.14 | 5.71 | 5.14 | 5.71 | 63.14 | 5.71 |
Mean | 11.39 | 10.36 | 8.29 | 7.14 | 6.49 | 6.66 | 6.53 | 6.92 | 6.74 | 5.14 | 53.97 | 19.39 |
For local image descriptors, a higher group of convolution layers can provide more discriminative descriptors than a lower group of convolution layers. Part of the multigroup union image descriptors can provide better resolution, and different arrangements of image descriptors from the same groups also have an impact on the results. The best average classification effect of the local image descriptors on the synthetic dataset was the multi-group union image descriptors composed of c3, c4, and c5, which was 93.08%. Local image descriptors achieved the best average classification effect on the measured dataset c5, which was 93.51%. The best classification effects of the local image descriptor on the synthetic and measured datasets were 100.00% and 99.71%, respectively.
For the global image descriptors, 𝒇1, which is transferred from the first fully connected layer, has strong semantic information, achieving an average accuracy of 92.76% and 94.86% on the synthetic dataset and measured dataset, respectively. The best classification effects of the global image descriptors on the synthetic and measured datasets were both 100.00%. In contrast, the classification model trained on the global image descriptors transferred from the second fully connected layer achieved a poor classification performance. The aforementioned experiments have preliminarily proved that image descriptors transferred from a DCNN model can be applied as the image descriptors of radionuclide gamma-ray energy spectrum images for classification; however, further comparisons of the features of gamma-ray energy spectra and different image descriptors of the gamma-ray energy spectrum images are needed.
Statistical comparison
Characteristic peaks are the most important features of traditional radionuclide identification methods. Four groups of characteristic peaks were extracted from the gamma-ray energy spectra by changing the peak properties, that is, distance, prominence, width, and threshold. Scale-invariant feature transform (SIFT)[50] and a histogram of oriented gradients (HOG)[51] are two classical feature extraction algorithms used in computer vision. Two groups of features extracted from the gamma-ray energy spectrum images were obtained using the HOG and SIFT algorithms. Fig. 7 demonstrates various features of the synthetic and measured data sets in a low-dimensional space through the t-SNE[48], which including characteristic peaks, features extracted by HOG and SIFT, present local image descriptors 𝒄5 and global image descriptors 𝒇1. As shown in Fig. 7, the distribution of the characteristic peaks is chaotic, and no apparent clustering center is observed for the different radionuclides. While HOG and SIFT features can form apparent clustering centers in certain radionuclides, the features of these radionuclides exhibit a significant crossover. Simultaneously, the transferred local and global image descriptors were significantly enhanced in feature discrimination, and the aggregation between similar feature points was also stronger.
The aforementioned features from one- and two-dimensional spectral images were applied in 21 classification methods using the Weka machine learning toolkit[49]. All experiments applied the ten-fold cross-validation method to avoid the imbalance caused by random data segmentation. The proportions of misclassified samples in the experiments are listed in Table 8 and Table 9. These results provide a basis for evaluating the performance of the transferred features. However, from a statistical point of view, these results do not provide strong support for identifying a group of features. For a statistical comparison of features, the non-parametric test method recommended by Demšar[52] was considered for comparing multiple features using several classification methods. First, we tested whether there were significant differences among the seven groups of features. The Friedman test was then applied to compare several features of the multiple classification methods.
Classification Method | peaks-1 | peaks-2 | peaks-3 | peaks-4 | 𝒇1 | HOG | SIFT |
---|---|---|---|---|---|---|---|
Bayes net | 0.0111 | 0.0236 | 0.0125 | 0.0332 | 0.0043 | 0.1091 | 0.4663 |
RBF network | 0.0625 | 0.0409 | 0.0721 | 0.0841 | 0.0000 | 0.2404 | 0.4976 |
Simple logistic | 0.0163 | 0.0899 | 0.1212 | 0.1986 | 0.0000 | 0.1562 | 0.4952 |
SMO | 0.1370 | 0.3389 | 0.3144 | 0.3649 | 0.0000 | 0.1562 | 0.4663 |
IBk | 0.0163 | 0.3389 | 0.0264 | 0.0538 | 0.0000 | 0.1774 | 0.4832 |
Bagging | 0.0139 | 0.0139 | 0.0298 | 0.0269 | 0.0024 | 0.1457 | 0.3918 |
Multi class classifier | 0.0308 | 0.2043 | 0.1625 | 0.2904 | 0.0024 | 0.4452 | 0.7572 |
JRip | 0.0163 | 0.0130 | 0.0274 | 0.0341 | 0.0019 | 0.2144 | 0.5673 |
J48 | 0.0087 | 0.0091 | 0.0216 | 0.0192 | 0.0120 | 0.1712 | 0.4784 |
LMT | 0.0101 | 0.0423 | 0.0221 | 0.0596 | 0.0000 | 0.1683 | 0.4952 |
Random forest | 0.0034 | 0.0043 | 0.0091 | 0.0087 | 0.0000 | 0.1187 | 0.3822 |
Mean | 0.0308 | 0.0726 | 0.0745 | 0.1067 | 0.0021 | 0.1912 | 0.4982 |
Classification Method | peaks-1 | peaks-2 | peaks-3 | peaks-4 | 𝒇1 | HOG | SIFT |
---|---|---|---|---|---|---|---|
Bayes net | 0.0943 | 0.1343 | 0.1114 | 0.1457 | 0.0486 | 0.1257 | 0.1257 |
RBF network | 0.2143 | 0.2857 | 0.1771 | 0.2600 | 0.0143 | 0.1314 | 0.1429 |
Simple logistic | 0.1057 | 0.1057 | 0.1343 | 0.1343 | 0.0057 | 0.1286 | 0.1629 |
SMO | 0.1971 | 0.1114 | 0.1457 | 0.1114 | 0.0057 | 0.1286 | 0.1286 |
IBk | 0.0914 | 0.1143 | 0.1371 | 0.0971 | 0.0000 | 0.1571 | 0.1343 |
Bagging | 0.1286 | 0.1143 | 0.1343 | 0.1343 | 0.0200 | 0.1457 | 0.1429 |
Multi class Classifier | 0.0943 | 0.1514 | 0.1571 | 0.1343 | 0.0200 | 0.1629 | 0.1600 |
JRip | 0.1514 | 0.1200 | 0.1229 | 0.1286 | 0.0600 | 0.1571 | 0.1400 |
J48 | 0.1314 | 0.0743 | 0.1171 | 0.1143 | 0.0257 | 0.1857 | 0.1714 |
LMT | 0.0857 | 0.1057 | 0.1371 | 0.1257 | 0.0057 | 0.1286 | 0.1714 |
Random forest | 0.0829 | 0.0800 | 0.0914 | 0.1057 | 0.0057 | 0.1171 | 0.1171 |
Mean | 0.0943 | 0.1557 | 0.1354 | 0.1356 | 0.0192 | 0.1426 | 0.1452 |
Assuming that Num types of classification methods and k groups of features are involved in this experiment, the implementation of the Friedman test assigns a value
In our experiments, the significance level α value was set to 0.05, the corresponding critical value of dataset 1 was calculated as 2.2541 while
The test statistics of the Holm test for comparing features in a pair-wise manner are formulated in Eq. (15).
If the adjusted α corresponding to the feature, α/(k-1), is higher than p1, the corresponding null hypothesis is rejected and indicates that both features have the same performance, and subsequently proceeds to compare p2 with the corresponding adjusted α, α/(k-2). If the second null hypothesis remains rejected, the experiment continues to test the third null hypothesis, and so on. Once a certain null hypothesis is accepted, all remaining null hypotheses are preserved.
In our case, 𝒇1 was chosen as the representative of the transferred image descriptors and compared with other groups of features. For data set 1,
Dataset name | i | Feature name | p | α/(k-i) | |
---|---|---|---|---|---|
Dataset1 | 1 | SIFT | (7.0000-1.1818) / 0.9211 = 6.3163 | 0.0000 | 0.0083 |
2 | HOG | (5.5455-1.1818) / 0.9211 = 4.7373 | 0.0000 | 0.0100 | |
3 | peaks-4 | (4.7727-1.1818) / 0.9211 = 3.8984 | 0.0000 | 0.0125 | |
4 | peaks-3 | (3.9091-1.1818) / 0.9211 = 2.9608 | 0.0031 | 0.0167 | |
5 | peaks-2 | (3.3636-1.1818) / 0.9211 = 2.3686 | 0.0179 | 0.0250 | |
6 | peaks-1 | (2.0909-1.1818) / 0.9211 = 0.9869 | 0.3237 | 0.0500 | |
Dataset2 | 1 | HOG | (5.3636-1.0000) / 0.9211 = 4.7373 | 0.0000 | 0.0083 |
2 | SIFT | (5.2273-1.0000) / 0.9211 = 4.5892 | 0.0000 | 0.0100 | |
3 | peaks-3 | (4.5000-1.0000) / 0.9211 = 3.7997 | 0.0001 | 0.0125 | |
4 | peaks-4 | (4.1818-1.0000) / 0.9211 = 3.4542 | 0.0006 | 0.0167 | |
5 | peaks-1 | (3.5455-1.0000) / 0.9211 = 2.7634 | 0.0057 | 0.0250 | |
6 | peaks-2 | (3.2273-1.0000) / 0.9211 = 2.4180 | 0.0156 | 0.0500 |
For Dataset 1, the Holm procedure rejects the first, second, third, fourth, and fifth hypotheses because the corresponding p values are smaller than the adjusted α. Thus, the final hypothesis cannot be rejected. This indicates that the transferred image descriptors perform significantly better than the characteristic peaks searched by prominence, width, and threshold, and features extracted by HOG and SIFT at the significance level of α=0.05. The transferred image descriptors were not significantly better than the characteristic peaks searched for by distance. For Dataset 2, the Holm procedure rejects all the hypotheses, which indicates that the transferred image descriptors perform significantly better than the characteristic peaks and features extracted by HOG and SIFT at the significance level α = 0.05. The filtering basis of searching the characteristic peaks by using the distance removes peaks with smaller distances by defining the minimum horizontal distance between adjacent peaks until all remaining peaks satisfy the distance condition. This peak-finding method is effective for the synthesized gamma-ray energy spectra using the NaI detector and has a larger classification error in the measured gamma-ray energy spectra using the CZT detector. Because the resolution of the NaI detector is not excellent, and the synthetic energy spectra comes from an ideal environmental simulation, the simulated energy spectrum has fewer interference peaks, while the measured gamma-ray energy spectra fluctuate significantly.
The results of the aforementioned comparative experiments prove that image descriptors transferred from DCNNs are better than the characteristic peaks extracted from the gamma-ray energy spectra and the HOG and SIFT shallow features extracted from the gamma-ray energy spectrum images. These transferred image descriptors are essential, discriminative, and can provide a reliable performance for gamma-ray energy spectrum image classification problems.
Conclusion
This study proposes a novel feature extraction approach for radionuclide identification to facilitate the extraction of structural and essential features and increase the precision of identification on the gamma-ray energy spectrum set.
The results of a series of comparative experiments between the proposed method, peak searching-based method, HOG, and SIFT using both synthetic and measured data demonstrate the following conclusions. (1) The information pre-processing of the proposed method, that is, mapping the gamma-ray energy spectra from a Euclidean space to a Banach space and then to an RGB color space, is significant for extracting the essential features, which can reduce the intraclass differences caused by different measuring durations. (2) The feature transfer process of the proposed method, that is, transferring the corresponding activation vectors of fully connected layers and activations from convolution layers in the process from DCNNs as image descriptors, can effectively extract the essential features of gamma-ray energy spectrum images. (3) Local image descriptors transferred from higher convolution layers provide more discriminative descriptors. (4) The global image descriptors transferred from the first fully connected layer had the strongest semantic information among the fully connected layers. (5) The proposed method outperforms the peak searching-based method, HOG, and SIFT on synthetic and measured datasets.
Future studies will focus on exploring the available value of other DCNNs in the field of radionuclide identification, exploring more feature fusion methods and aggregation approaches to develop more powerful descriptors, and establishing more universal datasets to further advance the research process of radionuclide identification.
Data for training and testing radiation detection algorithms in an urban environment
. Sci. Data. 7(1), 1-6 (2020). doi: 10.6084/m9.figshare.12654065A method for automatic identification of peaks in the presence of background and its application to spectrum analysis
. Nucl. Instrum. Methods. 50, 309-320 (1967). doi: 10.1016/0029-554X(67)90058-4fast nuclide identification based on sequential Bayesian method
. Nucl. Sci. Tech. 32(12), 143 (2021). doi: 10.1007/s41365-021-00982-zIdentification of unknown shielding parameters with gamma-ray spectra using a derivative-free inverse radiation transport model
. Nucl. Sci. Tech. 29(5), 70 (2018). doi: 10.1007/s41365-018-0401-5Estimation method for parameters of overlapping nuclear pulse signals
. Nucl. Sci. Tech. 28(1), 12 (2017). doi: 10.1007/s41365-016-0161-zUnfolding analysis of the LaBr 3: Ce gamma spectrum with a detector response matrix constructing algorithm based on energy-resolution calibration
. Nucl. Sci. Tech. 29, 1 (2018). doi: 10.1007/s41365-017-0340-6Statistical methods applied to gamma-ray spectroscopy algorithms in nuclear security missions
. Appl. Radiat. Isot. 70, 2428-2439 (2012). doi: 10.1016/j.apradiso.2012.06.016The application of neural networks for the analysis of gamma-ray spectra was measured with a Ge spectrometer
. Nucl. Instrum. Meth. A. 484(1-3), 557-563 (2002). doi: 10.1016/S0168-9002(01)01962-3Methods for obtaining the characteristic c-ray net peak count from the interlaced overlap peak in the HPGe c-ray spectrometer system
. Nucl. Sci. Tech. 30, 11 (2019). doi: 10.1007/s41365-018-0525-7Determination of net area for 92.6 keV peak of 238 U by the spectrum-stripping method
. Nucl. Tech. 15(4), 205-207 (1992). (in Chinese)Study on the Analysis Method of γ-Spectral Peaks of Environmental Samples
. J. Ningde Nor. Univ (Nat. Sci.). 13(2), 15-17 (1995). (in Chinese)In-situ gamma-ray spectrometer for the deep ocean
. Appl. Radiat. Isot. 142, 120-127 (2018). doi: 10.1016/j.apradiso.2018.08.024Real-time radionuclide identification in c-emitter mixtures based on spiking neural network
. Appl. Radiat. Isot. 109, 405-409 (2016). doi: 10.1016/j.apradiso.2015.12.029In situ spectrometry in a marine environment using full spectrum analysis for natural radionuclides
. Appl. Radiat. Isot. 114, 76-86 (2016). doi: 10.1016/j.apradiso.2016.05.008Nuclide identification algorithm based on K-L transform and neural networks
. Nucl. Instrum. Meth. A. 598(2), 450-453 (2009). doi: 10.1016/j.nima.2008.09.035Nuclide identification and analysis using artificial neural networks
. Ord. Indu. Auto. 34(11),86-91(2015) doi: 10.7690/bgzdh.2015.11.022Radioactive Nuclide Identification Method Based on SVD and SVM
. Ord. Indu. Auto. 36(5),50-53 (2017). (in Chinese)Wavelet analysis of gamma-ray spectral
. Paper Presented at the 13th IEEE Symposium Conference Record on Nuclear Science (Wavelet analysis of sodium iodide spectral
. IEEE T. Nucl. Sci. 53(5), 2916-2922 (2006). doi: 10.1109/TNS.2006.881909a rapid radionuclide identification algorithm based on the discrete cosine transform and a BP neural network
. Ann. Nucl. Energy. 112, 1-8 (2018). doi: 10.1016/j.anucene.2017.09.032nuclide spectrum feature extraction, and nuclide identification based on a sparse representation
. High Power Laser Part. Beams. 30(04), 046003 (2018). doi: 10.11884/HPLPB201830.170435Radionuclide identification algorithm for an organic scintillator-based radiation portal monitor
. Nucl. Instrum. Meth. A. 849, 41-48 (2017). doi: 10.1016/j.nima.2017.01.009Expectation propagation for weak radionuclide identification at radiation portal monitors
. Sci. Rep.-Uk. 10(1), 1-12 (2020). doi: 10.1038/s41598-020-62947-3Comparative study of machine learning methods for automated identification of radionuclides using NaI gamma-ray spectra
. Nucl. Eng. Technol. 53(12), 4072-4079 (2021). doi: 10.1016/j.net.2021.06.020Radionuclide identification method based on a gamma-spectra template library synthetic by radial basis function neural networks
. J. Tsinghua Univ (Sci. and Tec.). 61(11), 1308-1315 (2021). doi: 10.16511/j.cnki.qhdxxb.2020.22.033Automated isotope identification algorithm using artificial neural networks
. IEEE T. Nucl. Sci. 64, 1858-1864 (2017). doi: 10.1109/TNS.2017.2693152Validation of a Bayesian-based isotope identification algorithm
. Nucl. Instrum. Meth. Phys. Res. Sect. A. 784, 298-305 (2015). doi: 10.1016/j.nima.2014.11.113Investigation of fuzzy recognition mechanism for c-ray fingerprints of nuclear materials
. Acta Phys. Sinica. 57(9), 5361-5365 (2008). doi: 10.7498/aps.57.5361Radionuclide identification method for NaI low-count gamma-ray spectra using an ANN
. Nucl. Eng. Technol. 54(1), 269-274 (2022). doi: 10.1016/j.net.2021.07.025L.H. Tsoukalas. Kernel-based machine learning for background estimation of NaI low-count gamma-ray spectra
. IEEE. T. Nucl. Sci. 60(3), 2209-2221(2013). doi: 10.1109/TNS.2013.2260868Intelligent analysis of low-count scintillation spectra using support vector regression and fuzzy logic
. Nucl. Technol. 191(1), 41-57 (2015). doi: 10.13182/NT14-75The study of nuclide identification algorithm based on sequential Bayesian analysis
. Nucl. Elec. Det. Tech. 36(2), 179-183 (2016). doi: 10.3969/j.issn.0258-0934.2016.02.015 (in Chinese)Energy spectrum nuclide recognition method based on long short-term memory neural networks
. High Power Laser Part. Beams 32(10), 106001 (2020). doi: 10.11884/HPLPB202032.200118 (in Chinese)Development of a radionuclide identification algorithm based on a convolutional neural network for a radiation portal monitoring system
. Radiat. Phys. Chem. 180, 109300 (2021). doi: 10.1016/j.radphyschem.2020.109300Applying deep-learning in gamma-spectroscopy for radionuclide identification
. Paper Presented at the 3rd International Conference on Dosimetry and its Applications. (Transferring DCNNs for scene classification of high-resolution remote sensing imagery
. Remote Sens-Basel. 7(11), 14680-14707 (2015). doi: 10.3390/rs71114680Neural codes for image retrieval
. Paper Presented at the 13th European Conference on Computer Vision. (Multiscale orderless pooling of deep convolution activation features
. Paper Presented at the 13th European Conference on Computer Vision. (visual instance retrieval using deep convolutional networks
. ITE Trans. Media Technol. Appl. 4(3), 251-258 (2016). doi: 10.3169/mta.4.251Deep filter banks for texture recognition and segmentation
. Paper Presented at the 28th IEEE Conference on Computer Vision and Pattern Recognition. (Aggregating local deep features for image retrieval
. Paper Presented at the 15th IEEE International Conference on Computer Vision. (Aggregating deep convolution features for image retrieval
(2015). arXiv:1510.07493The treasure beneath convolution layers: Cross-convolution-layer pooling for image classification
. Paper Presented at the 28th IEEE Conference on Computer Vision and Pattern Recognition. (A New Colormap for MATLAB - Part 1 - Introduction
, https://blogs.mathworks.com/steve/2014/10/13/a-new-colormap-for-matlab-part-1-introduction/; 2014 [accessedThe homepage of ImageNet
, https://ImageNet.org/; 2021 [accessedObject recognition from local scale-invariant features
. Paper Presented at the 7th IEEE International Conference on Computer Vision. (Histograms of oriented gradients for human detection
. Paper Presented at the 18th IEEE Computer Society Conference on Computer Vision and Pattern Recognition. (Very deep convolutional networks for large-scale image recognition
. Paper Presented at the 3rd International Conference on Learning Representations (Visualizing data using t-SNE
. J. Mach. Learn. Res. 9(11),2579-2605 (2008). https://lvdmaaten.github.io/publications/papers/JMLR_2008.pdfStatistical comparisons of classifiers over multiple data sets
. J. Mach. Learn. Res. 7, 1-30 (2006).