Introduction
UAV-borne gamma-ray spectrum survey is to measure and record the type and intensity of gamma-ray emitted by natural radionuclides (e.g., uranium, thorium, potassium) on the ground and shallow surfaces. This data, combined with the coordinate information, is processed, screened and corrected to produce maps of surface radionuclide distribution, essential for geological exploration, radioactive mineral prospecting, and environmental monitoring. Compared to fixed-wing aircraft, UAV-borne gamma-ray spectrum surveys offer low flight costs, high measurement efficiency, safe and flexible flight processes, and prolonged hovering capabilities. Consequently, they have been widely adopted in radioactive mineral exploration, radiation environment monitoring, nuclear emergency response, and related fields [1-3].
Zhang [4] skillfully used tension spline method and Munsell transform technology to carry out fine rasterization and anomaly correction of aerial gamma spectrum data, successfully extracted key information, clearly displayed the energy spectrum characteristics of Xiangshan volcanic basin, and analyzed the distribution of uranium, thorium and potassium, providing new clues for uranium exploration and promoting the development of geological exploration technology. Šálek [5] tested the performance of a new type of small airborne gamma ray spectrum measurement equipment based on UAV, and verified its ability to detect uranium anomalies at different altitudes, and can accurately capture small changes in gamma ray intensity even at higher altitude. Li [6] developed a curviline-based noise reduction technique for airborne gamma data, which can reconstruct multi-scale data for different analysis purposes, effectively remove noise from the data, while retaining local anomaly information to avoid resolution loss and boundary effects. Lyu [7] proposed a K-factor prediction model considering flight altitude and direct distance. The model pays special attention to the key factors such as path loss, shadow fading and small-scale fading at different flight altitudes, which brings potential application value to the field of aerial gamma-ray spectral image processing. Wang [8] proposes a layered approach specifically for REM recovery of limited sampled data in unknown environments, demonstrating the possibility of achieving high-precision REM construction under low sampling rates. The data recovery algorithm and sampling optimization strategy can be used for reference in the research of aerial gamma-ray spectral image processing to improve the processing accuracy and efficiency.
The demand for high-resolution UAV-borne gamma-ray spectrum images has increased to improve the accuracy of research. For instance, in sandstone-type uranium exploration and radiation environment monitoring, high-resolution images allow researchers to identify weak anomalies and locate radionuclide distributions more accurately. However, the direct acquisition of ultra-high-resolution gamma-ray spectrum images is challenging due to UAV load and flight altitude limitations.
Image Super-Resolution (SR) technology, which converts low-resolution (LR) images into high-resolution (HR) images, can significantly enhance the details and data information of existing images. This technology plays an important role in many fields such as satellite imagery [9], face recognition [10], and medical imaging [11]. In recent years, with the development of deep learning, people begin to use deep learning technology to improve image super resolution. The SR method based on convolutional neural network, residual network (ResNet) and generative adversarial networks (GAN) are proposed. SRCNN [12], as the first SR deep learning network, achieves fast online application with its lightweight structure and excellent recovery quality, despite its large computational load. ESPCN [13] introduces an innovative subpixel convolution layer to achieve LR to HR super-resolution at a minimal computational cost. VDSR [14] uses a very deep convolutional network, combined with residual learning, to significantly improve the convergence speed during training. These methods improve the accuracy and speed of image super resolution by using faster and deeper convolutional neural networks. However, when enlarging the LR image to the HR image, the reconstructed SR image often lacks texture details due to the large scale factor, resulting in unsatisfactory reconstruction effect. SRGAN [15-17] enhanced the content loss function with counter-loss by training GAN, and replaced the content loss based on Mean Squared Error (MSE) with the loss based on VGG network feature map, effectively overcoming the problem of low perception quality of reconstructed images and making the generated images more realistic. However, the downside of this method lies in its complex network structure and lengthy training process.
However, degradation of image quality is a complex phenomenon, usually due to the limitations of the imaging system and environmental disturbances. Blind super resolution (Blind SR) technology arises at the historic moment. It can effectively restore LR images to HR images when the degradation process is not completely clear. This technology can be divided into two categories: display modeling and implicit modeling. Display modeling by parameterizing fuzzy kernel and noise information. SRMD [18-20], as the first deep learning-driven Blind SR method, introduces a dimension expansion strategy, which enables the convolution network to handle fuzzy kernel and noise level inputs, and solves the dimension mismatch problem, although it may perform poorly for uncovered degradation types. The BSRGAN model [21] proposed by Zhang et al. in 2021 enhances the adaptability of models to real-world degradation by introducing complex degradation factors and random shuffling strategies. Based on the kernel stretching strategy of Luo [22] proposed a new practical degradation model, which uses dynamic depth linear filter and constrained least square deconvolution algorithm based on neural network to improve the restoration quality of blurred images. Implicit modeling abandons explicit parameters and uses deep learning techniques, especially GAN, to restore directly from LR images to HR images. The CinCGANcite [23, 24] model adopts double cyclic GAN structure, which effectively solves the problem of complex interference in low resolution input. The research classification of Liu [25] points out that some methods train SR models by learning the degradation process from HR to LR and using the generated LR samples, such as Degradation GAN [26], FSSR [27] and FSSRGAN [28], but these methods may have domain gap problems. DASR improves the performance of SR training through domain gap awareness training and domain distance weighted monitoring strategy. Real-ESRGAN [29] improved on the basis of ESRGAN [30], using discriminator designed by U-Net and high-order degradation process, and introducing sinc filter to reduce ringing and overshoot artifacts, providing a more accurate and stable super-resolution solution for real-world images.
In this paper, a new scheme combining UAV-borne gamma-ray spectrum survey and image super-resolution technology is proposed to overcome the limitations of existing technology in obtaining high-resolution images. Real-ESRGAN technology can improve the clarity and feature enhancement of UAV-borne gamma-ray spectrum images, and then help to interpret geological data more accurately. For example, it can enhance areas of the image that are blurred due to topographic effects, making geological features more clearly visible. In addition, Real-ESRGAN’s de-noising and detail enhancement capabilities help to remove noise generated during flight measurements, thereby improving data quality. Image super resolution technology can improve LR image to HR through algorithm processing, and compensate for the shortage of hardware equipment. This technology, combined with the UAV-borne gamma-ray spectrum survey, can not only improve the clarity and detail expression of the image, but also effectively improve the accuracy and reliability of data analysis. Therefore, the combination of UAV-borne gamma-ray spectrum survey technology and image super-resolution technology can more effectively monitor environmental radiation levels, but also provide more accurate data support for the efficient exploration of radioactive minerals, and promote the development of radioactive geophysical exploration technology.
Technical Principles and Methods
Technical Principles and Methods
UAV-borne gamma-ray spectrum technology captures gamma rays of different energy levels from the ground and shallow surfaces using a gamma detector mounted on the UAV. This technology is promising for geological exploration and radiation environment monitoring. However, the images generated often suffer from noise and disturbances, leading to potential misinterpretations of geological information.
In the field of image processing, a series of noise filtering and image enhancement techniques are usually needed to improve the quality of UAV-borne gamma-ray spectrum images. The Gaussian filter is widely used, which can reduce the interference of noise to the image quality by smoothing the image. The formula of Gaussian filtering [31] can be expressed as:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M001.png)
Median filtering [32] is also a common nonlinear filtering technique, which reduces noise by replacing the value of a pixel with the median value in its neighborhood. For a given image I, the formula for calculating the pixel value of the image I’ after median filtering at (x, y) is:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M002.png)
Although this method surpasses Gaussian filtering in preserving edge information, it may cause boundary blurring when processing images with high contrast edges. To further emphasize geological features in the AGS images, image enhancement techniques such as contrast enhancement are used to improve the dynamic range of the images, making subtle details more visible. Edge detection technology can identify boundary lines in the image, which is particularly critical for determining the precise location of geological structures. However, these techniques may increase the artificial distortion of images, leading to over-emphasis or misidentification of geological features.
Image Super-Resolution Reconstruction
Image Super-Resolution aims to convert LR images into HR images by modeling the degradation process, which includes blurring, downsampling, and noise. The goal is to find an operator that makes the reconstructed image as close as possible to the original HR image. Under the Bayesian framework, this process is expressed as a Maximum A Posteriori (MAP) problem [33], which involves a least squares problem with a regularization term to incorporate prior knowledge. The core goal of image super resolution reconstruction technology is to convert LR images into HR images. Mathematically, this process is often modeled as a degradation process where the LR image is regarded as the result of the HR image being blurred, downsampled, and disturbed by noise._2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M003.png)
The goal of super-resolution reconstruction is to find an operator R, which makes _2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M004.png)
Under the Bayesian framework [34], super-resolution reconstruction can be expressed as a MAP problem._2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M005.png)
In practice, since it is usually not feasible to compute _2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M006.png)
In the field of deep learning, Real-ESRGAN and other algorithms approximate the optimal reconstruction operator R* by training a depth neural network. At the same time, the GAN is used to improve the visual quality and naturalness of the reconstructed image.
Super-Resolution Reconstruction of Gamma-Ray Spectrum Image of UAV
Real-ESRGAN Principle
ESRGAN (Enhanced Super-Resolution Generative Adversarial Networks), based on GAN, generates high-quality super-resolution images through competition between generator and discriminator. Real-ESRGAN improves upon ESRGAN specifically for the super-resolution reconstruction of real-world images. By introducing higher-order degradation models and Sinc filters, it can more accurately simulates the image degradation process in the real world, including blur, downsampling, noise and JPEG compression among others. Additionally, Real-ESRGAN employs a U-Net structure discriminator and spectral normalization technology, which not only enhances the discriminator’s ability to distinguish, but also improves the stability of the training process.
(1) GAN
GAN(Generative Adversarial Networks) [35] consists of two main components: Generator and Discriminator. As shown in Fig. 1. The generator is responsible for converting the low-resolution image into a high-resolution image, while the discriminator is responsible for distinguishing between the generated high-resolution image and the real high-resolution image. The two are trained through a process of adversarial interaction, where the generator continually refines its output to deceive the discriminator, while the discriminator concurrently enhances its capabilities to more effectively distinguish between generated and real images.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F001.jpg)
(2) Network architecture
Real-ESRGAN retains the Residual-in-Residual Dense Block (RRDB) [36] from ESRGAN as the core component of its generator. Through an innovative network structure design, the performance of image super-resolution reconstruction is significantly improved. The RRDB is composed of multiple residual blocks, each of which is further connected in a dense manner, allowing features from all previous layers to be directly connected to the current layer. RRDB enhances the learning capability of the network through multiple residual connections while simultaneously eschewing the use of Batch Normalization (BN) layers, which contributes to the improvement of the detail clarity in the generated images. In addition, the RRDB’s role in a generator is multifaceted. It not only acts as the backbone for feature extraction, but also works collaboratively with other parts of the network through residual connections, such as sampling layers and feature fusion modules, to generate high-quality high-resolution images. Figure 2 shows the structure of the RRDB. By this design, Real-ESRGAN is able to produce detailed and visually realistic super-resolution images to meet the needs of a variety of practical applications.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F002.jpg)
(3) Loss function
To achieve high-quality reconstruction, Real-ESRGAN employs a composite loss function, a design that ensures the accuracy of the generated high-resolution images at the pixel level as well as visual fidelity and richness of detail. The compound loss function consists of the following key components:
1) L1 loss (Mean Absolute Error, MAE)
L1 loss measures the average of the absolute difference between the predicted value and the true value, and its mathematical expression is defined as:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M007.png)
2) Perceptual Loss
Perceptual loss is typically based on pre-trained convolutional neural networks (such as VGG networks) to extract features and compare the differences in these features between the generated and real images. A simplified perceived loss can be expressed as:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M008.png)
3) Adversarial Loss
By training a discriminator to distinguish between a real image and a generated image, the generator’s goal is to maximize the probability that the discriminator will make a wrong judgment. A common form of countering loss is the use of Wasserstein distance, which is expressed as:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M009.png)
The loss functions synergistically work together to ensure that the generated high-resolution images exhibit not only fidelity at the pixel level, but also visually realistic and detailed enough to meet the needs of a variety of real-world applications.
Data source
The data in this paper are from a experimental area in northern Gansu province, China, which is shown in Fig. 3. A total of 62889 data point from total count (Tc), uranium (U), thorium (Th) and potassium (K) channels were measured and processed to create equivalent maps of surface radionuclide distribution. To enhance the visual quality and detail resolution of UAV-borne gamma-ray spectrum images, image restoration techniques were employed on the raw data in this study. Consequently, the color scales in the UAV-borne gamma-ray spectrum images presented herein are dimensionless, serving solely to enhance contrast and visualization rather than for quantitative measurement of radioactive count rates. In our study, in order to adapt to the specific resolution and characteristics of the UAV-borne gamma-ray spectrum data, we performed segmentation and selection on different regions of the original UAV-borne gamma-ray spectrum images, directly skipping areas where the cropping region exceeds the image boundaries, ultimately forming 232 two-dimensional slices of 80×80 pixels each. The 80×80 pixel cutting size was chosen to take full advantage of the gamma-ray detector’s spatial resolution at the actual flight altitude, while maintaining enough detail for effective geological feature analysis.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F003.jpg)
In order to obtain low resolution UAV-borne gamma-ray spectrum images, the ‘resize’ function from the ‘PIL.Image’ library was utilized to alter the image size by specifying new dimension parameters (width and height). The ‘resample=Image.LANCZOS’ parameter specifies the use of the ‘Lanczos’ resample algorithm, which is a high quality resampling method suitable for image scaling. As shown in Fig. 4, with the increase of the UAV-borne gamma spectrum image to 6 times, some geological detail loss can be clearly observed.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F004.jpg)
Model training and analysis
The hardware used includes an INTEL I5-13600KF 14-Core processor with 3.5GHz, 32GB memory and an NVIDA GeForce RTX 4070Ti graphics card. The RealESRGAN model was constructed using PyTorch framework, and the Adam optimizer was used to train the model at a learning rate of 1×10-4, exponential motion average (EMA) was employed for more stable training. In addition, L1 loss, perceived loss and GAN loss were combined for training with weights of (1,1,0.1), respectively. L1 loss ensures the accuracy of the reconstructed image at the pixel level, perceptual loss is enhances the high-level visual quality of the image to align with human perception, and GAN loss increases the realism and naturalness of the image. The comprehensive use of different types of loss functions to improve the effect of UAV-borne gamma-ray spectrum super-resolution image reconstruction.
UAV-borne gamma-ray spectrum images are a special type of remote sensing image, providing important data for geological exploration, mineral resource development and environmental monitoring by detecting and recording the distribution of radioactive elements on the surface. These images need high spatial resolution to capture subtle geological features. The main advantage of super-resolution image is their ability to significantly enhance the spatial resolution of the image, making the visual effect of the image more clearer, detailed and more realistic. To evaluate the application of super-resolution reconstruction technology in UAV-borne gamma-ray spectrum images, this study selected the Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) as the main evaluation indicators.
PSNR is a crucial metric for measuring image quality, primarily used to evaluate the similarity between the reconstructed image and the original image, providing a quantitative analysis of pixel-level errors between images. Its unit is decibels (dB), a higher value, indicates that the reconstructed image is to the original image, thereby reflecting better image quality. The calculation of PSNR depends on MSE, the formula is:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M010.png)
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M011.png)
SSIM is an index used to measure the similarity between two images. It evaluates the image quality through brightness, contrast and structure to reflect the perceived quality of the image more comprehensively. Therefore, the calculation of SSIM is also based on three main components: brightness, contrast and structure. For the original image x and the reconstructed image y, the SSIM index is:_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M012.png)
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-M013.png)
Results and Discussion
In this paper, we reconstructed the UAV-borne gamma-ray spectrum images with 2×, 4× and 6× super resolution. SRCNN algorithm, SRGAN algorithm, FSRCNN algorithm and Real-ESRGAN algorithm are compared to verify the effectiveness of Real-ESRGAN algorithm.
To validate the model performance, this study randomly selected 10 images from the image set of four elements: U, Th, K, and Tc, to constitute a test set. During the testing process, Th and K were primarily chosen for analysis. By comparing and analyzing with the geological map of the survey area (as shown in Fig. 3), it was found that the distribution patterns of Th and K within the survey area exhibit a high degree of consistency. Notably, there is a fault within the survey area that extends from the northwest to the southeast, which aligns closely with the high-value bands of Thorium and Potassium. This phenomenon is attributed to the distribution of Hercynian late-stage granites, which have caused an anomaly enrichment of radioactive Potassium and Thorium in survey area. Typically, granites contain minerals such as potassium feldspar and mica, which are rich in Potassium; concurrently, they also contain Thorium-bearing minerals like monazite and xenotime. Consequently, in areas where granite is distributed, the content of Thorium and Potassium is generally high. Based on this, the test dataset employed in this paper can authentically reflect the distribution of natural radioactive nuclides within the survey area.
To visually demonstrate the superior effect of the Real-RESGAN algorithm on the super-resolution reconstruction of UAV-borne gamma-ray spectrum images, the three introduced super-resolution reconstruction algorithms were applied with 2×, 4× and 6× super-resolution on the test set. The results are shown in Figs. 5, 6, and 7.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F005.jpg)
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F006.jpg)
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F007.jpg)
The analyses delineated within Figs. 5, 6 and 7 expose discernible disparities in the efficacy of assorted algorithms engaged in the reconstruction of imagery. The super-resolution reconstructions yielded by the SRCNN algorithm exhibit a measure of obscurity and chromatic aberration, with a pronounced exacerbation in blurriness concomitant to escalated magnification. Particularly, at the 6× magnification exemplified in Fig. 7, the loss of detail information leads to the obscurity of geological body boundaries, rendering them difficult to discern. The FSRCNN algorithm, an advanced rendition of its progenitor, ameliorates the acuity of high-frequency detail capture through the refinement of convolutional strata and activation mechanisms, thereby bestowing a marked enhancement in the textural, marginal, and configurational fidelity of the reconstructed imagery relative to the SRCNN (as depicted in Fig. 5). Despite these advancements, discrepancies persist between the FSRCNN’s output and the authentic geological background and boundaries, with residual haziness observable (as illustrated in Fig. 7). The SRGAN algorithm surpasses its precursors in the holistic quality of image reconstruction, integrating the GAN architecture to facilitate a more nuanced rendition of intricate image details. Even with the magnification increased to 6×, the detailed information and boundaries of the geological bodies remain distinctly visible, with a significant reduction in the sense of image blurriness. In stark contrast to SRCNN, FSRCNN, and SRGAN, the Real-ESRGAN algorithm demonstrates a pronounced superiority in reconstructed image fidelity, efficaciously attenuating artifacts such as ringing and overshoot. As portrayed in Fig. 5, the algorithm enables the distinct identification of depositional loci across macroscopic and microscopic scales. The reconstructed imagery not only elucidates the ore body’s continuity but also accentuates the demarcation between the ore body and the enveloping rock matrix, affording an exact visual portrayal of the ore deposit’s morphology, magnitude, and orientation. However, an increase in magnification intensity introduces an over-saturation at the ore body periphery, as noted in Fig. 7.
In order to comprehensively verify the superiority of Real-ESRGAN algorithm, this study further selected and reconstructed U, Th, K and Tc images in the test set, as shown in Fig. 8. The results show that the image reconstructed by the Real-ESRGAN algorithm has a significant improvement in overall sharpness and color vividness. A smoother and more natural transition is displayed at the edge of the ore body and key geological landmarks, effectively avoiding the misunderstanding and error that may arise in the image processing. This not only improves the overall resolution of the image, but more importantly, enhances the detailed representation of geological structures and deposit sites, providing more abundant and accurate identifiable features for geological exploration and resource evaluation.
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F008.jpg)
When evaluating the quality of super-resolution reconstructed images, in addition to the subjective analysis based on visual senses, this study also adopted a variety of objective quantitative indicators to verify the accuracy of experimental results. Not only the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) are introduced as evaluation criteria, but the error between the reconstructed image and the original image is objectively quantified, as shown in Table 1 and Table 2. Residual analysis between the original image and the generated super-resolution image has also been added by calculating the residual image between the two and analyzing its histogram, the ability of Real-ESRGAN algorithm in detail recovery and noise processing is revealed, which provides a comprehensive perspective for evaluating the quality of super-resolution reconstructed images as shown in Fig. 9.
| Magnification | SRCNN | SRGAN | FSRCNN | Real-ESRGAN |
|---|---|---|---|---|
| 2 | 16.914 | 26.176 | 20.215 | 29.176 |
| 4 | 15.167 | 25.109 | 16.942 | 26.408 |
| 6 | 12.907 | 24.326 | 15.177 | 23.827 |
| Magnification | SRCNN | SRGAN | FSRCNN | Real-ESRGAN |
|---|---|---|---|---|
| 2 | 0.659 | 0.934 | 0.802 | 0.950 |
| 4 | 0.502 | 0.906 | 0.638 | 0.942 |
| 6 | 0.272 | 0.868 | 0.502 | 0.938 |
_2026_02/1001-8042-2026-02-29/alternativeImage/1001-8042-2026-02-29-F009.jpg)
The analysis results from Table 1 and Table 2 show that with the increase of magnification, the PSNR and SSIM values of reconstructed images obtained by each algorithm generally show a downward trend. Notably, the Real-ESRGAN algorithm demonstrates a pronounced superiority in both indices, signifying its exceptional performance in reconstructing super-resolution images. When juxtaposing these quantitative measures, it is evident that the Real-ESRGAN algorithm’s reconstructions are markedly closer to the original images, particularly observed in the SSIM values, which excel in capturing the structural fidelity and perceptual quality of the imagery. The SSIM metric, with its sensitivity to both local alterations and the granularity of textures, provides a comprehensive assessment of image quality. Illustratively, at a 2× magnification, the Real-ESRGAN algorithm achieves an SSIM value of 0.950, a Figure that underscores the close resemblance of the reconstructed image to its pristine counterpart. This high SSIM value not only attests to the Real-ESRGAN algorithm’s proficiency in preserving fine details but also substantiates its overall efficacy and preeminence in the realm of super-resolution image reconstruction tasks.
It can be seen from Fig. 9 that the residual values are mainly concentrated in the central position, and most of the residual values are close to zero. This shows that the difference between the super-resolution reconstructed image and the original image is very small on most pixels, and the Real-ESRGAN model can better retain the distribution characteristics of radioactive elements in the original image. Although most of the residuals are concentrated in the center, it can also be seen that some residuals are too large, which may be due to the differences between the reconstructed image and the original image of the geological boundary, high-frequency structure or noisy area. The symmetry of the histogram reflects that the residuals are evenly distributed in positive and negative directions. This symmetry means that the Real-ESRGAN model has no obvious systematic bias and can restore the radiation intensity change of the UAV-borne gamma-ray spectrum image in a balanced way.
Conclusion
In this study, the Real ESRGAN algorithm is successfully applied to reconstruct the UAV-borne gamma-ray spectrum image in super resolution, significantly enhancing the spatial resolution of the image and improving the visualization quality of geological features. By comparing with SRCNN, SRGAN and FSRCNN algorithms, Real-ESRGAN has shown excellent performance in PSNR and SSIM objective evaluation indices. Especially, under the condition of 2× amplification, the SSIM value is as high as 0.950, which substantiates its advantage in detail preservation and texture clarity, and further highlights the significant enhancement in the identification of geological body boundaries. Additionally, the Real-ESRGAN algorithm effectively reduces ringing and overshoot artifacts, making the transition between the edge of the ore body and key geological markers more smooth and natural, and greatly enhancing the detailed representation of the geological body structure and deposit sites. This clear delineation of lithological boundaries provides geologists with more intuitive and accurate geological information, thereby offering significant application value in geological exploration and resource assessment. Consequently, the Real-ESRGAN algorithm is not only theoretically advanced but also demonstrates robust practical utility, providing an effective image processing tool for UAV-borne gamma-ray spectrum image processing.
Mapping mafic dyke swarms, structural features, and hydrothermal alteration zones in Atar, Ahmeyim and Chami areas (Reguibat Shield, Northern Mauritania) using high-resolution aeromagnetic and gamma-ray spectrometry data
. J. Afr. Earth Sci. 163,Radioactive mineralization detection using Remote Sens-Basel and airborne gamma-ray spectrometry at Wadi Al-Miyah area, Central Eastern Desert, Egypt
. The Egyptian Journal of Remote Sens-Basel and Space Sciences. 25, 37–53 2022). https://doi.org/10.1016/j.ejrs.2021.12.004Airborne HPGe spectrometer for monitoring of air dose rates and surface activities
. Nucl. Eng. Technol. 55, 4039–4047 2023). https://doi.org/10.1016/j.net.2023.07.019P. Soc Photo-Opt Ins and analysis of airborne spectral data in Xiangshan uranium ore field
. Journal of East China University of Technology: Natural Science Edition 35(02), 124–128 2012). https://doi.org/CNKI:SUN:HDDZ.0.2012-02-005Mapping of radiation anomalies using UAV mini-airborne gamma-ray spectrometry
. J. Environ. Radioact. 182, 101–107 2018). https://doi.org/10.1016/j.jenvrad.2017.11.033Research on the application of curvelet transform in airborne gamma spectral Data Process-London
. World Nuclear Geoscience 40(4), 963–973 (2023). https://doi.org/10.3969/j.issn.1672-0636.2023.04.007Fixed-Wing UAV based air-to-ground channel measurement and modeling at 2.7GHz in rural environment
. IEEE T. Antenn. Propag. 73, 2038–2052 2025). https://doi.org/10.1109/TAP.2024.3428337Sparse bayesian learning-based hierarchical construction for 3D radio environment maps incorporating channel shadowing
. IEEE T. Wirel. Commun. 23, 14560–14574 2024). https://doi.org/10.1109/TWC.2024.3416447Frequency-Assisted mamba for Remote Sens-Basel image super-resolution
. 2024. https://doi.org/10.48550/arXiv.2405.04964Feature super-resolution based facial expression recognition for multi-scale low-resolution images
. Knowl-Based Syst. 236,Simultaneous Tri-Modal medical image fusion and super-resolution using conditional diffusion model
. 2024. https://doi.org/10.48550/arXiv.2404.17357Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network
, in 2016 IEEE ConferencSe on Computer Vision and Comm Com Inf Sc (CVPR).Accurate image super-resolution using very deep convolutional networks
, in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Photo-realistic single image super-resolution using a generative adversarial network
, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Image super-resolution reconstruction based on SRGAN
. Computer Knowledge and Technology 20(01), 14–17 2024). https://doi.org/10.14004/j.cnki.ckt.2024.0029Research on the super-resolution of engineering cost bills based on SRGAN
. Intelligent City 9(10), 105–107 2023). https://doi.org/10.19301/j.cnki.zncs.2023.10.033Learning a single convolutional super-resolution network for multiple degradations
, in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (Research on adaptive image super-resolution algorithms
.Research on blind image super-resolution algorithms based on deep learning
.Designing a practical degradation model for deep blind image super-resolution
. 2021. https://doi.org/10.48550/arXiv.2103.1S4006Deep constrained least squares for blind image super-resolution
. 2022. https://doi.org/10.48550/arXiv.2202.07508An innovative application of generative adversarial networks for physically accurate rock images with an unprecedented field of View
. Geophys. Res. Lett. 47(23),Blind image super-resolution: A survey and beyond
. IEEE T. Pattern Anal. 45, 5461–5480 2022). https://doi.org/10.1109/TPAMI.2022.3203009To learn image super-resolution, use a GAN to learn how to do image degradation first
, in European Conference on Computer Vision. (Frequency separation for realworld super-resolution
, in 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) (Guided frequency separation network for real-world super-resolution
, in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (Real-esrgan: training real-world blind super-resolution with pure synthetic data
. 2021. https://doi.org/10.48550/arXiv.2107.10833Esrgan: enhanced super-resolution generative adversarial networks
. 2018. https://doi.org/10.1007/978-3-030-11021-5_5Gaussian filtering of images: A regularization approach
. Signal Process 18(2), 169–181 1989). https://doi.org/10.1016/0165-1684(89)90048-0On detection of median filtering in digital images
. International Society for Optics and Photonics 7541,Three-dimensional maximum a posteriori (MAP) imaging with radiopharmaceuticals labeled with three Cu radionuclides
. Nucl. Med. Biol. 33(2), 217–226 2006). https://doi.org/10.1016/j.nucmedbio.2005.11.001MEG source localization under multiple constraints: an extended bayesian framework
. Neuroimage 30(3), 753–767 (2007). https://doi.org/10.1016/j.neuroimage.2005.10.037Image generation using generative adversarial networks and attention mechanism
, in 2016 IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS) (CT super-resolution using multiple dense residual block based GAN
. SIViP 15, 725–733 2021). https://doi.org/10.1007/s11760-020-01790-5The authors declare that they have no competing interests.

