Introduction
Traditional x-ray computed tomography (CT) employs an energy-integrating detector whose output is proportional to the energy integrated over the entire incident spectrum, while spectral responses of different materials are ignored [1,2]. To address this issue, photon-counting CT uses an energy-resolving detector to gain energy-dependent information, which can be applied to multi material detection and decomposition [3,4]. In classical material decomposition methods, the attenuation of objects is usually expressed as a linear combination, such as a linear combination of photoelectric absorption, incoherent (Compton) and coherent (Rayleigh) scattering [5,6] or several basis materials [7]. However, owing to the low signal-to-noise ratio of spectral CT images in a narrow energy bin, the material decomposition performance is affected by artifacts and noise, particularly for material components with similar x-ray absorption characteristics.
Several methods have been proposed to solve these problems. These methods can be classified into three categories: one-step inversion, projection domain-based, and image domain-based methods. One-step inversion algorithms combine reconstruction and decomposition processes to realize material decomposition directly from projection data. Mory et al. investigated the convergence speed of five one-step algorithms. One-step inversion methods are typically slower than their two-step counterparts because of their high computational cost [8]. Tilley et al. presented a model-based material decomposition algorithm that enabled reconstruction and decomposition to be synchronously performed using a multienergy forward model [9]. Meanwhile, Fang et al. proposed an iterative one-step inversion material decomposition algorithm with a Noise2Noise prior. This algorithm estimated material images directly from projection data and used a Noise2Noise prior for denoising [10].
Projection domain-based methods first decompose projection data into sinograms of materials and then perform reconstruction to obtain material images. Noh et al. proposed two novel sinogram restoration methods based on statistical models and suppress noise [11]. Petrongolo et al. used convex optimization to reduce noise in decomposed images in the projection domain [12]. Image domain-based methods first reconstruct images and then perform decomposition in the image domain. Zhao et al. developed an image-domain material decomposition algorithm by incorporating an edge-preserving filter into a local highly constrained backprojection reconstruction framework [13]. Niu et al. formulated an algorithm that included the inverse of the estimated variance-covariance matrix of the decomposed images as a penalty weight in the least-square term [14]. Chen et al. attempted to use the end-to-end AUTOMAP for fitting solutions in the image domain [15]. By using simulated data, Clark et al. trained CNNs with a U-Net structure [16]. After training, they tested the CNNs on experimentally acquired dual energy-energy integrating detector- and photon counting detector-based micro-CT data in mice. Their U-Net performed a more robust spectral micro-CT decomposition compared to sensitivity-based decomposition [17]. Chen et al. proposed a convolutional material decomposition algorithm that exhibited robustness during multimaterial decomposition of spectral CT images even when obvious ring artifacts were present [18]. Zimmerman et al. separated the soft tissue, bone, and gadolinium nanoparticles of an ex vivo rat leg specimen into different basis images with machine learning and transfer learning [19].
Theoretically, the purpose of spectral CT material decomposition in the image domain is to decompose the different material components in spectral CT images, which is the same goal as that of image segmentation tasks performed in certain clinical applications. Lungs, bones, and tissues are treated as the base material. Recently, deep learning has been widely used for image segmentation [20], classification [21], object detection [22], image denoising [23, 24], and artifact reduction [25, 26]. Deep learning can be used to perform feature learning and hierarchical feature extraction to solve image segmentation problems; thus, it is feasible to use it improve the accuracy of material decomposition of spectral CT images. Currently, some relevant studies have used deep learning technology for material decomposition. Feng et al. designed an end-to-end network model to obtain a virtual monochromatic attenuation map from a multicolor reconstruction map for material decomposition [27]. Xu et al. proposed a deep neural network composed of a fully convolutional network (FCN) and a fully connected network to solve a material decomposition problem [28]. Wu et al. presented a fully convolutional DenseNet (FC-DenseNet) based on densely connected blocks of Densenet to classify and quantify different materials [29]. Zhang et al. studied a model-based butterfly network to perform image domain material decomposition for dual-energy CT [30]. Gong et al. developed a neural network called Incept-net, which had an encoder—decoder framework, for material decomposition of spectral CT images [31]. Inspired by the generative adversarial network, Geng et al. proposed a novel parallel multi-stream generative adversarial network (PMS-GAN) to perform projection-based multi-material decomposition in spectral CT, in which they designed multiple sub-generators to perform multimaterial decomposition simultaneously [32]. Popescu et al. applied Pix2Pix GAN in retinal images and achieved satisfactory results in the segmentation of blood vessels; their network included a U-Net type generator and PatchGAN type discriminator [33]. However, these two GAN-based studies did not apply an attention mechanism and ignored some key feature information. Ma et al. proposed a residual dense convolutional neural network for low-dose CT denoising via an attention mechanism, which achieved excellent performance [34]. Lan et al. introduced a self-attention mechanism to a conditional generative adversarial network, which enabled robust deep learning-based neuroimaging synthesis. The self-attention module calculated the similarity among pixels of the image [35], which is computationally expensive. Unlike the methods proposed in previous studies, our global attention mechanism can extract global information using considerably fewer parameters.
To effectively improve the accuracy of material identification based on deep learning, we introduce the concept of adversarial learning to material decomposition research [36,37] and propose an attention-based global convolutional generative adversarial network (AGC-GAN) for material decomposition, which is capable of extracting feature information at multiple scales of spectral CT images. AGC-GAN employs large convolutional kernels to extract global feature information in feature maps at multi-scales and embeds an attention mechanism [38] in the fusion process of multi-layers of feature maps. Spectral CT images datasets were constructed by scanning a mouse specimen with a photon-counting CT system. Then, we performed experiments to evaluate the performance of AGC-GAN on this spectral CT image dataset. Compared to U-Net, FCN [39], and FCD [40] our proposed method achieved a better discrimination performance on three materials: lung, bone, and soft tissue.
Methodology
OVERVIEW OF AGC-GAN
Although there exist similar structural features of spectral CT images among different energy bins, they contain different gray feature information. Therefore, it is possible to construct a sufficiently large dataset according to the gray differences and structural correlation of spectral CT images and to improve the accuracy of material decomposition in spectral CT images by deep learning technology. Deep learning can be used for feature information extraction and learning, allowing inherent laws and deep features of the available dataset to be learnt.
In this study, we use deep learning technology to identify lung, bone, and soft tissue from the chest spectral CT images of a mouse. Generative adversarial networks (GANs) have recently become a powerful model in the deep learning field, opening a new avenue for deep learning frameworks. Based on the GAN algorithm, we designed AGC-GAN for material decomposition of spectral CT images. The structure of AGC-GAN is shown in Fig. 1. Our network consists of two sub-modules: a generator module and a discriminator module. First, spectral CT images are fed into the generator network to produce the corresponding material decomposition prediction maps. Then, the discriminator network uses the reference image and the prediction map to determine the authenticity. The generator produces material decomposition images with the help of feedback from the discriminator and deceives the discriminator, while the discriminator attempts to distinguish between the generative images and the reference images. Functionally, the adversarial learning process corrects inconsistencies between predicted decomposition maps and reference images.
-202303/1001-8042-34-03-013/alternativeImage/1001-8042-34-03-013-F001.jpg)
NETWORK STRUCTURE
GENERATOR NETWORK
As shown in Fig. 1, AGC-GAN has an attention-based global convolutional network (AGCN), which is a U-type network, as the generator. Resnet-152 [41] was employed as the backbone network for AGCN in this study, and the details of backbone are presented in Table 1. Initially, multi-scale feature maps of inputs were extracted from different stages in the backbone network. Subsequently, a global convolutional block (GCB) was employed to extract the global information of the feature map, and a boundary refinement block (BRB) was used to help the model learn the boundary information of the feature map [42]. The GCB uses a large convolution kernel to achieve global convolution, enabling the feature information at each level to be closely related to the pixel categories. However, the large kernel size convolution operation entails excessive computation and storage requirements during the learning process. Therefore, GCB adopts parallel asymmetric convolution to complete the large-kernel-size convolution, effectively reducing the number of parameters and computational burden of the network.
Conv2d (kerne_size, out_channels) | |
---|---|
Resnet152-layer0 | 7×7, 64, stride 2 |
Resnet152-layer1 | 3×3 max pool, stride 2 |
Resnet152-layer2 | |
Resnet152-layer3 | |
Resnet152-layer4 |
Low-resolution feature maps were then up-sampled by using an up-sampling block (USB) and merged with higher-resolution feature maps to generate new feature maps. The final prediction maps were generated after the last BRB, which was used to output the decomposition results of AGCN. The compositions of the GCB, BRB, and USB are shown in Table 2.
GCB | BRB | USB |
---|---|---|
1×15 Conv, 15×1 Conv | 3×3 Conv | 3×3 Transposed Conv |
15×1 Conv, 1×15 Conv | BN + ReLu 3×3 Conv |
To enhance the feature extraction performance of the generator network, we added an attention mechanism to the fusion of feature maps for attention mapping. The attention mechanism can help the network assign different weights to various parts of feature maps and extract critical information, enabling it to make more accurate judgments in the learning process. Recently, the attention mechanism has become a powerful tool for performing deep learning tasks. Woo et al. proposed the convolutional block attention module (CBAM), which improves the accuracy of networks by concatenating spatial and channel attention [43]. Inspired by the CBAM, we designed a channel attention block (CAB) and a spatial attention block (SAB) and incorporated them into the generator network. Figure 2 shows the structure of the CAB and SAB.
-202303/1001-8042-34-03-013/alternativeImage/1001-8042-34-03-013-F002.jpg)
The CAB contributes to assigning different weights for convolutional feature maps. These weights indicate the importance of channel-wise feature maps. Specifically, two different layers of feature maps provide the inputs of the CAB. First, feature maps of two different scales are concatenated in the channel dimension and then fed into 1×1 convolution (Conv), batch normalization (BN), and rectified linear unit (ReLU) layers sequentially after an average pooling (Avg-Pool) operation. Second, the channel attention map is obtained using the sigmoid activation function. Finally, the input feature map of the CAB is dot multiplied (DM) with the channel attention map and summed with the next-level feature map to obtain the feature information extracted by the CAB as the output. The SAB helps the network assign different weights to various spatial regions according to their feature content. Specifically, the SAB takes the output of the CAB as its input. First, the input feature maps are subjected to maximum pooling (Max-Pool) and Avg-Pool; subsequently, they are concatenated in the channel dimension. The feature map is then sequentially fed into 3×3 Conv, BN, and ReLU layers and activated using the sigmoid function to obtain the spatial attention map. Eventually, the input feature map is DM with this spatial attention map to obtain the final output feature map. Thus, more informative and discriminative features can be obtained using the CAB and SAB, thereby effectively enhancing the decomposition performance of the network.
DISCRIMINATOR NETWORK
The discriminator network of AGC-GAN in this study is PatchGAN, which is also called a Markovian discriminator [44]. As shown in Fig. 1, the original spectral CT image and the corresponding decomposition result image are concatenated together as the input of discriminator. Specifically, the reference image is considered “real” for the discriminator, while the decomposition result image produced by the generator is considered “fake”. The discriminator mainly consists of 4 down blocks (DB) and 1 output convolutional layer that finally outputs a 30×30 size confidence map. The structural parameters of the discriminator network are listed in Table 3.
Down sampling block 1 | 4×4 Conv LeakyReLU |
Down sampling block 2 |
LOSS FUNCTION
The loss of GAN based on binary cross-entropy loss calculated at the discriminator can be defined as
In this study, the objective loss function of the AGC-GAN designed herein consists of two components, the generative adversarial loss and the cross-entropy loss of the generator network, which is defined as shown in Eq. (2):
SPECTRAL CT DATASET
To obtain our spectral CT image dataset, we used a spectral CT imaging system based on a photon counting detector to scan a mouse specimen in our study. The photon-counting detector (SANTIS 0804) produced by DECTRIS has four fully calibrated and independently adjustable energy gating thresholds, and its effective detection area is 1024×256 pixels, with a pixel size is 150 μM. The SANTIS 0804 is a hybrid photon counting (HPC) detector with two configurations: high-resolution mode and multi-energy mode. We set the detector mode of the spectral CT imaging system to multi-energy mode, where the maximum input photon counting rate is 4.0×108 photons/s/mm2.
We selected one mouse of approximately 10 cm in length and 160-180 g in weight to be scanned. The scanned mouse specimen was treated with the required reagents. The scanning parameters for the mouse specimen in the experiment are listed in Table 4. The spectral CT system scanned the mouse three times, and the SANTIS 0804 detector was set with four different energy thresholds for each scan. Consequently, we obtained spectrum data of the mouse specimen in 12 different energy bins. We selected 10 distinctive spectral ranges from 12 energy channels, namely 25–90, 30–90, 35–90, 40–90, 45–90, 50–90, 55–90, 60–90, 65–90, and 70–90 keV. In each scanning process, 250 projection views were acquired for each energy bin. We selected 100 slices from the projection data to reconstruct CT images, and each spectral CT image mainly contained three materials: lung, bone, and soft tissue. The selected projection data was reconstructed via the Split-Bregman algorithm, and finally a dataset containing 1000 spectral CT images was constructed.
Tube voltage (kvp) | 90 |
Tube current (μA) | 200 |
Source to detector distance (mm) | 350 |
Source to mouse specimen distance (mm) | 210 |
Total number of angels | 250 |
Length of detector | 1024 |
Number of projections | 256 |
In addition, we manually made labels for each spectral CT image by using LabelMe, a professional labeling software. Each label image had the same resolution as the original spectral CT image. Some of the original spectral CT images and labeled reference images in the final dataset are shown in Fig. 3. Because all the data were from one mouse, we performed 5-fold cross-validation experiments to avoid overfitting. We divided the 100 slices into five equal parts from mouse head to mouse tail. Every part contained 20 slices, with every slice having 10 images with 10 different energy bins. We used three parts for training, one part for validation, and the remaining part for testing. We repeated this process five times as shown in Fig. 4 and reported the average results on these five testing parts to evaluate the performance of the model.
-202303/1001-8042-34-03-013/alternativeImage/1001-8042-34-03-013-F003.jpg)
-202303/1001-8042-34-03-013/alternativeImage/1001-8042-34-03-013-F004.jpg)
EXPERIMENTAL RESULTS
To verify the feasibility of AGC-GAN for material decomposition of spectral CT images, U-Net, FCN, and FCD were selected for comparison. The programming language used in the experiments was Python 3.7, and the individual network was implemented based on the deep learning framework PyTorch 1.5.1. The hardware platform configuration for running the network is as follows: Intel i5-9600kf CPU, Nvidia TITAN V (12GB/Nvidia) GPU, and 16G DDR4 RAM. The initialization learning rate of each network was set to 0.001, the decay rate of the learning rate was 0.95/(2 epochs), and the parameters of each layer of network were optimized by the Adam (Adaptive Moment Estimation) algorithm. In every rotation, we used the same training dataset to train AGC-GAN and other comparison networks, and we then used the same test dataset to test the decomposition performance of each network. The differences among the results of each network are shown in Fig. 5. The results of the three slices are from rotation 1, 3, and 4.
-202303/1001-8042-34-03-013/alternativeImage/1001-8042-34-03-013-F005.jpg)
From Fig. 5, we can see that the decomposition results of AGC-GAN are closer to the ground truth compared with other networks. AGC-GAN achieves superior material decomposition results, particularly for lung and bones. In addition, to further evaluate the material decomposition performance of each network, we calculated intersection over union (IoU), structural similarity (SSIM) index, precision, pixel accuracy (PAcc), and F1-score (F1) metrics for each network. IoU is a standard measurement method to check the degree of similarity of the output with the ground truth. IoU can be expressed as
SSIM index is used as a metric to measure the similarity between two given images.
Networks | IoU | |||
---|---|---|---|---|
Lung (%) | Bone (%) | Tissue (%) | MIoU (%) | |
U-Net | 86.84±3.32 | 80.27±4.28 | 94.09±0.60 | 87.20±2.48 |
FCN | 86.57±3.88 | 79.00±4.46 | 93.65±0.82 | 86.54±2.86 |
FCD | 79.45±8.89 | 58.59±6.98 | 90.98±1.76 | 76.46±5.38 |
GCN-GAN | 85.52±3.60 | 75.30±3.99 | 93.71±1.09 | 84.84±2.64 |
AGC-GAN | 87.00±3.01 | 80.37±3.21 | 94.54±1.08 | 87.31±2.16 |
Networks | SSIM | Precision | |||
---|---|---|---|---|---|
Lung (%) | Bone (%) | Tissue (%) | MPre (%) | ||
U-Net | 94.21+0.53 | 93.37+2.29 | 88.04+5.35 | 97.04+0.91 | 92.95+1.77 |
FCN | 93.96+0.62 | 93.80+1.72 | 87.83+4.91 | 96.94+1.00 | 92.89+2.00 |
FCD | 92.37+1.11 | 89.01+3.64 | 66.24+8.93 | 95.71+0.63 | 83.65+4.20 |
GCN-GAN | 93.32+0.74 | 93.39+3.09 | 82.39+3.61 | 96.74+1.21 | 90.84+1.69 |
AGC-GAN | 94.83+0.81 | 93.85+1.19 | 88.68+1.87 | 97.13+0.99 | 93.22+0.85 |
Networks | PAcc | F1-score | |||
---|---|---|---|---|---|
Lung (%) | Bone (%) | Tissue (%) | Mean (%) | ||
U-Net | 97.21±0.41 | 92.48±0.19 | 88.05±2.75 | 96.48±0.32 | 92.47±1.53 |
FCN | 97.18±0.42 | 92.31±2.25 | 87.34±2.83 | 96.25±0.43 | 92.10±1.73 |
FCD | 96.32±1.76 | 88.21±1.31 | 73.54±5.59 | 95.26±0.97 | 86.42±5.11 |
GCN-GAN | 97.35±0.87 | 92.12±2.13 | 85.72±2.67 | 96.75±0.58 | 91.53±1.65 |
AGC-GAN | 97.39±0.47 | 93.00±1.75 | 88.97±2.00 | 97.19±0.58 | 93.05±1.29 |
DISCUSSION AND CONCLUSION
Photon-counting CT is a promising new technology for CT systems. It can provide spectral information for identifying different components than existing technologies, and it can improve the accuracy of material identification. However, the noise in images in narrow energy bins is very strong, making the grey value of images inaccurate. Traditional material decomposition methods do not provide satisfactory decomposition results. Although spectral CT images with different energy bins have different grey values, they have the same structural similarity. Finally, we proposed a novel methodology to improve the accuracy of material decomposition. We intend to replace traditional material decomposition with segmentation based on deep learning. This will help improve the accuracy and aid radiologists in diagnosis. Our global attention mechanism contains channel attention and spatial attention. For photon-counting CT data, every slice has ten images under ten different energy bins. Channel attention mechanism can assign different weights to every channel, while spatial attention mechanism assigns different weights on the same position of every channel. Thus, we can extract feature information of the different energy bins. We also utilized the global convolution block and boundary refinement block to realize global convolution. However, self-attention mechanism usually uses a three-feature space for calculating the similarity among image pixels, which is computationally expensive and requires more samples to avoid overfitting. We performed five-fold cross-validation experiments to test our algorithm. The proposed method achieved the best values for precision, IoU, PAcc, and F1-score metrics in comparison with three other networks, providing better performance for the material decomposition of spectral CT images. Note that the proposed AGC-GAN here is used for material decomposition of two-dimensional tomographic images. There is a high structural similarity and a strong correlation in the spatial structure for spectral CT images with different energy ranges. Therefore, we will use volume data to train a three-dimensional network model in our follow-up research. This model could learn more local feature information in the spatial dimension, improving the accuracy of material decomposition.
In summary, we proposed a material decomposition method, named AGC-GAN, for spectral CT images. In our network, the generator generates the results of material decomposition, while the discriminator attempts to distinguish whether the inputs are the reference or generated image. The generator and discriminator facilitate the material decomposition accuracy of AGC-GAN by learning against each other in a confrontational manner. The experimental results demonstrated that compared with U-Net, FCN, and FCD, our network exhibited excellent results the material decomposition of spectral CT. We performed cross-validation experiments to avoid the overfitting problem. In the future, a large amount of photon-counting CT data will be collected to improve the generalization ability of our network.
Photoncounting CT: technical principles and clinical prospects
. Radiology 289, 293 (2018). doi: 10.1148/radiol.2018172656Optimization of K-edge imaging with spectral CT
. Med. Phys. 39, 6572 (2012). doi: 10.1118/1.4754587Tutorial on X-ray photon counting detector characterization
. J. X-Ray Sci. Technol. 26, 1-28 (2018). doi: 10.3233/XST-16210Multi-material decomposition of spectral CT images
, in Proc. SPIE 7622, Medical Imaging 2010: Physics of Medical Imaging, 7622w ( 2010). doi: 10.1117/12.844531Energy-selective reconstructions in X-ray computerised tomography
. Phys. Med. Biol. 21, 733 (1976). doi: 10.1088/0031-9155/21/5/002Virtual monoenergetic CT imaging via deep learning
. Patterns 1, 100128 (2020). doi: 10.1016/j.patter.2020.100128Generalized image combinations in dual KVP digital radiography
. Med. Phys. 8, 659 (1981). doi: 10.1118/1.595025Comparison of five onestep reconstruction algorithms for spectral CT
. Phys. Med. Biol. 63, 235001 (2018). doi: 10.1088/1361-6560/aaeaf2Model-based material decomposition with a penalized nonlinear least-squares CT reconstruction algorithm
. Phys. Med. Biol. 64, 035005 (2019). doi: 10.1088/1361-6560/aaf973Iterative material decomposition for spectral CT using self-supervised Noise2Noise prior
. Phys. Med. Biol. 66, 155013 (2021). doi: 10.1088/1361-6560/ac0afdStatistical sinogram restoration in dual-energy CT for PET attenuation correction
. IEEE Trans. Med. Imaging 28, 1688 (2009). doi: 10.1109/TMI.2009.2018283A general framework of noise suppression in material decomposition for dual-energy CT
. Med. Phys. 42, 4848 (2015). doi: 10.1118/1.4926780Using edge-preserving algorithm with non-local mean for significantly improved image-domain material decomposition in dual-energy CT
. Phys. Med. Biol. 61, 1332 (2016). doi: 10.1088/0031-9155/61/3/1332Iterative image-domain decomposition for dual-energy CT
. Med. Phys. 41, 041901 (2014). doi: 10.1118/1.4866386Material decomposition of energy spectral CT by AUTOMAP
. IEEE Nucl. Sci. Conf. Proc. (2018). doi: 10.1109/NSSMIC.2018.8824439U-Net: convolutional networks for biomedical image segmentation
, in Lect. Notes. Comput. SC. (2015), p. 234–241. doi: 10.1007/978-3-319-24574-4_28Multi-energy CT decomposition using convolutional neural networks
. Proc. SPIE 10573, 1057310 (2018). doi: 10.1117/12.2293728Robust multimaterial decomposition of spectral CT using convolutional neural networks
. Opt. Eng. 58, 013104 (2019). doi: 10.1117/1.OE.58.1.013104Experimental investigation of neural network estimator and transfer learning techniques for K-edge spectral CT imaging
. Med. Phys. 47, 541 (2020). doi: 10.1002/mp.13946PCANet: a simple deep learning baseline for image classification? IEEE Trans
. Image Process. 24, 5017 (2015). doi: 10.1109/TIP.2015.2475625Image segmentation using deep learning: a survey
. IEEE Trans. Pattern Anal. (2021). doi: 10.1109/TPAMI.2021.3059968Object detection with deep learning: a review
. IEEE Trans. Neural Netw. Learn. Syst. 30, 3212 (2019). doi: 10.1109/TNNLS.2018.2876865Robust restoration of low-dose cerebral perfusion CT images using NCS-Unet
. Nucl. Sci. Tech. 33, 1-15 (2022). doi: 10.1007/s41365-022-01014-0DU-GAN: generative adversarial networks with dual-domain U-net based discriminators for low-dose CT denoising
. IEEE Trans. Instrum. Meas. 71, 4500512 (2022). doi: 10.1109/TIM.2021.3128703Study on analytical noise propagation in convolutional neural network methods used in computed tomography imaging
. Nucl. Sci. Tech. 33, 1-14 (2022). doi: 10.1007/s41365-022-01057-3Rollback reconstruction for TDC enhanced perfusion imaging
. Nucl. Sci. Tech. 32, 1-11 (2021). doi: 10.1007/s41365-021-00918-7A multi-energy material decomposition method for spectral CT using neural network
. Proc. SPIE 10573, 105734J (2018). doi: 10.1117/12.2294611Image decomposition algorithm for dual-energy computed tomography via fully convolutional network
. Comput. Math. Method Model. (2018). doi: 10.1155/2018/2527516Multi-material decomposition of spectral CT images via Fully Convolutional DenseNets
. J. X-Ray Sci. Technol. 27, 461 (2019). doi: 10.3233/XST-190500Image domain dual material decomposition for dual-energy CT using butterfly network
. Med. Phys. 46, 2037 (2019). doi: 10.1002/mp.13489Deep-learning-based direct inversion for material decomposition
. Med. Phys. 47, 6294 (2020). doi: 10.1002/mp.14523PMS-GAN: parallel multi-stream generative adversarial network for multi-material decomposition in spectral computed tomography
. IEEE Trans. Med. Imaging 49, 571-584 (2020). doi: 10.1109/TMI.2020.3031617Retinal blood vessel segmentation using Pix2Pix GAN
, in MED. (2021), p. 1173–1178. doi: 10.1109/MED51440.2021.9480169Sinogram denoising via attention residual dense convolutional neural network for low-dose computed tomography
. Nucl. Sci. Tech. 32, 41 (2021). doi: 10.1007/s41365-021-00874-2SC-GAN: 3D self-attention conditional GAN with spectral normalization for multi-modal neuroimaging synthesis
. bioRxiv (2020). doi: 10.1101/2020.06.09.143297Generative adversarial networks
. Adv. Neural Inf. Process. Syst. (2014). doi: 10.1145/3422622A transformer generative adversarial network for multi-track music generation
. CAAI Trans. Intell. Technol. 7, 369-380 (2022). doi: 10.1049/cit2.12065Attention-based novel neural network for mixed frequency data
. CAAI Trans. Intell. Technol. 6, 301-311 (2021). doi: 10.1049/cit2.12013Fully convolutional networks for semantic segmentation
, in Proc. CVPR. IEEE (2015), p. 3431–3440. doi: 10.1109/TPAMI.2016.2572683Fully convolutional pyramidal residual network for material discrimination of spectral CT
. IEEE Access 7, 167187 (2019). doi: 10.1109/ACCESS.2019.2953942Deep residual learning for image recognition
, in Proc. CVPR. IEEE (2016), p. 770–778. doi: 10.48550/arXiv.1512.03385Large kernel matters—improve semantic segmentation by global convolutional network
, in Proc. CVPR (CBAM: convolutional block attention module
, in ECCV (2018), p. 3–19. doi: 10.1007/978-3-030-01234-2_1Image-to-image translation with conditional adversarial networks
, in Proc. CVPR (