Figures
Abstract
Positron imaging has shown great potential in industrial non-destructive testing due to its high sensitivity and ability to reveal internal structures of complex components. However, reconstructing high-quality images from positron emission data remains challenging, particularly under limited sampling and ill-posed inverse problems, which are common in applications such as closed cavity detection. To address this, we propose an iterative reconstruction method for industrial positron images based on a generative adversarial network (PIIR-GAN). The method integrates a generative adversarial framework with a self-attention mechanism to exploit prior information and improve image quality under low-sample conditions. A key innovation is embedding the neural network model directly into the iterative reconstruction process, enabling end-to-end learning. Furthermore, a likelihood-based constraint is incorporated into the objective function to guide optimization. Experimental results on a GATE simulation dataset show significant improvements in both PSNR and SSIM compared with conventional methods, and real-world industrial defect detection further verifies the effectiveness of the approach.
Citation: Zhu M, Zhao M, Yao M (2025) Iterative reconstruction of industrial positron images with generative networks. PLoS One 20(11): e0335912. https://doi.org/10.1371/journal.pone.0335912
Editor: Hui Li, Dalian Maritime University, CHINA
Received: February 24, 2025; Accepted: October 17, 2025; Published: November 19, 2025
Copyright: © 2025 Zhu et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting information files.
Funding: National Natural Science Foundation of China (No. 62071229).
Competing interests: The authors have declared that no competing interests exist.
Introduction
Positron Emission Computed Tomography (PET) is functional imaging technology which is highly sensitive. Compared with other traditional industrial non-destructive testing method, such as X-ray and CT, the gamma photons produced in positron annihilation process have stronger penetrability and lower radiation. Therefore, it has a better application prospect in the detection of industrial airtight cavity with high precision.
The quality of PET image reconstruction largely depends on the number of response lines, which directly determines the amount of useful information. However, under current industrial conditions, available samples are limited, and the acquisition process often suffers from noise, sparse sampling, and environmental interference. Due to the ill-posed nature of the inverse problem, artifacts and noise frequently degrade the final image, reducing its utility in real-world industrial inspections.
Therefore, improving the PET imaging effect has become a current research focus for the application of PET technology in non-destructive testing of industrial defects. Existing research on PET image reconstruction mainly focuses on two aspects: one is processing the image in the post-reconstruction stage to enhance image quality [1–4]; the other is improving image quality by incorporating prior knowledge [5–8]. Methods of the latter type act directly on the reconstruction stage of positron images, better preserving the data obtained from sampling and thereby improving image quality. Although some studies on PET image reconstruction have made progress, most of these methods are applied to medical PET images. In industrial applications, however, due to constraints such as field environment, hardware limitations, and sampling efficiency, the performance of these methods in practical non-destructive testing remains unsatisfactory.
To address this issue, in the paper we discuss a reconstruction model for industrial positron images based on deep learning, building upon existing related studies. Specifically, we propose an approach to improve the quality of iterative reconstruction using Generative Adversarial Networks (GANs) [9] under low-sampling conditions in industrial settings. The method enables the acquisition of higher-quality defect detection images in complex industrial environments. Experimental results on both simulated and real images demonstrate that the proposed approach significantly improves image quality. Compared to existing deep learning-based PET reconstruction methods, which are primarily developed for medical imaging with relatively clean and abundant data, our approach is designed to handle the challenging conditions of industrial applications. Unlike typical post-processing or denoising methods, our model embeds prior knowledge directly into the iterative reconstruction process, preserving more sampling details and achieving higher accuracy under low-sample, high-noise conditions.
In this study, we discuss the mathematical model in the positron imaging process firstly. Then we use GAN to iteratively reconstruct the images. Finally, we conduct various experiments to determine the effects of adopting deep learning model in the reconstruction.
The main contributions of this paper are as follows.
- Get the prior input of the image reconstruction by training a small number of simulation data samples;
- Incorporated the deep neural network in the image iterative reconstruction stage to avoid the loss of sampling features.
- Optimize the neural network structure based on the combination of perceptual loss function and attention loss function.
The paper is organized as follows. The Related work section introduces the related work. The Method section describes the method we proposed in the paper. The Experiment section shows the experimental results and discusses the results in the discussion section. Finally, we provide conclusions in the last Section.
Related work
In recent years, the application of deep learning in the field of computer vision has made a lot of progress. Compared with traditional methods, the deep neural network model has obvious performance improvement in image classification, image denoising, object detection, semantic segmentation and etc. Meanwhile, deep neural networks, which used to reconstruct images is also a research hotpot in the field of imaging. For example, deep convolutional networks were used to adjust the coefficients of wavelet transform and reduced the reconstruction noise of low-dose CT images [10]. On the basis of the above framework, a wavelet residual network denoising algorithm is proposed, which has good performance in preserving image texture details [11]. [12] presented a residual network model which combine auto-encoder and deconvolution networks to realize the structure fidelity and noise suppression of low-dose CT images. In PET imaging, [13] proposed a reconstruction method using a dynamic convolutional module that maps low-dose PET plus CT images into standard-dose PET. [14] presented an unsupervised CNN that uses cross-multiplication regularization on list-mode data to improve positron image reconstruction accuracy. [15] introduced LegoPET, a conditional diffusion model guided by hierarchical features for PET reconstruction from sinograms, outperforming prior methods in PSNR/SSIM. [16] proposed DREAM, which integrates random masks in both sinogram and latent spaces to better capture both local details and global structures.
The most of the common models on deep neural networks are constructed based on large-scale training data. However, when PET is used for industrial non-destructive testing, the sample data obtained under current conditional is less. Therefore, in order to get better reconstructed images, GANs are considered in the process of image reconstruction. It is a deep learning method based on probability and statistics theory to generate data samples. It does not require or only needs less labeled data to learn from the idea of game theory for efficient data generation. The model consists of generator and discriminator. The generator generates data and the discriminator judges the true or false, then the error of the discriminant result is passed to the network to improve the parameters. The above process is repeated continuously to optimize the performance of the model until it reached the Nash equilibrium stat, and the mathematical expression is as Eq (1).
Given the strong potential of GANs in generative modeling, many works have adapted GANs for image reconstruction tasks. For instance, [17] added a latent code in the random noise input to control the data generation, and added a mutual information regular term to indicate the degree of association. [18] combined the advantages of SRGAN (Super-Resolution Generative Adversarial Network) [19] and RaGAN (Relativistic Average Generative Adversarial Network) [20]. It used residual dense block units and a relative average discriminator to make the edges of the reconstructed images are sharper. [21] used the general reconstruction loss, gradient loss and additional adversarial loss to train full convolution network, and it successfully synthesized high quality real images. [22] proposed a distributed network, the first one synthesizes images, and the second used image translation framework to obtain higher resolution. [23] built a dual channel generative network to get more realistic global output based on conditional generative adversarial network. [24] proposed a controllable GAN based on the combination with generator of ResNet-like and discriminator of PatchGAN [25] and realized the faithful reconstruction of images. GAN and autoencoder were combined for image reconstruction and the model trained positive samples to build the images based on using local binary pattern for image local contrast to detect defects [26]. [27] proposed a self-supervised adaptive residual GAN (SS-AEGAN) that mitigates texture inconsistencies and enhances low-dose PET image quality through adaptive residual mapping and self-supervised pretraining. [28] proposed PCC-GAN, a point-based GAN that enhances PET reconstruction quality by capturing geometric and contextual relationships from low-dose data.
At the same time, to obtain better model training effects, more model regularization and generalization researches are also being further developed in the field of deep learning. [29] proposed a spectrum interference-based two-level data augmentation method in deep learning for automatic modulation classification. [30] optimized the specifically designed autoencoder (AE) by entropy-stochastic gradient descent. [31] improved he pruning method by reducing the network parameters and the calculation cost.
In summary, a number of studies have been conducted to overcome the challenging in image reconstruction, however, several aspects thereof have been not yet been satisfactorily resolved. In this study, we focus on addressing the following problem of images reconstruction under industrial positron imaging conditions. (i) the data range of the sample is small, (ii) data limitation under limited conditions, (iii) image quality is not very satisfactory. To solve these limitations, we decided to employ a GAN-based iterative reconstruction approach to obtain higher quality PET images, particularly in a more complex industrial environment.
Method
PET data model
The process of producing γ photons by positron annihilation can be abstracted into a Poisson distribution model and the mean value depends on the distribution of radionuclides. The goal of PET image reconstruction is to obtain the nuclide position distribution function in relative space. The measured data can be modeled as a collection of independent Poisson distribution and its mean
is related to the image
through statistical iteration. The model can be abstracted Eq (2).
Where is the system matrix, in which Ai,j represents the photons originating from voxel j and detected by detector i.
denotes the expectation of scattered events.
denotes expectation of random coincidences. M is the number of lines of response (LOR) and N is the value of pixel in image space. The log-likelihood function can be written as Eq (3).
Reconstruction model
The reconstructed image x, which mentioned above is considered to be represented by the model as Eq (4).
Where denotes the neural network, and α is defined as the input to the adversarial network. By training the neural network with existing data, prior knowledge of industrial positron images is incorporated into the reconstruction framework.
And the maximum likelihood estimates of the images x is as Eq (5).
To improve reconstruction accuracy, we introduce an additional likelihood term to adjust the dimensionality of the random input α to match the feature scale of positron images. Then the maximum likelihood estimation of the image x can be calculated as Eq (6).
Here, the hyperparameter λ is used to balance the α and x to ensure the stability of the model and both of them are matched the sampled positron data.We constrain the network optimization with different values of λ, and after multiple training experiments, was found to achieve stable results.
Then we use the Augmented Lagrange method to optimize the constraint problem above and the expression is as Eq (7).
And this expression can be solved by ADMM (Alternating direction method of multipliers) algorithm as Eq (8).
where μ denotes the Lagrange multiplier introduced in the augmented Lagrangian formulation, which is iteratively updated in the ADMM optimization process to enforce the constraint .
The reconstruction framework can include the constraint through pre-trained adversarial network model using existing images.We adopt a GAN-based structure as the backbone network, considering its ability to learn low-dimensional manifolds of high-count samples and achieve better performance than conventional convolutional networks under limited sampling conditions.
We set the input to the generator is the low-count image sample and the input to the discriminator is the same resolution high-count image. The whole net is summarized in Fig 1. The generator consists of 4 up-sampling layers and 4 down-sampling layers, and the discriminator consists of 4 down-sampling layers.
The generative nets in our method are based on the U-net structure [32] and includes batch normalization layer. The attention model is also added in the nets which inspired by the attention U-net [33]. It consists of repetitive use of (1) 3×3 convolutional layers, (2) batch normalization layers, (3) relu layers, (4) convolutional layers with stride, (5) transposed convolutional layers with stride. And in the generative nets, the major modifications are as follows: (1) fully convolutional nets are used to implement pixel segmentation, (2) attention gate is added in the left-side layer to screen features, (3) residual network is used to contact the input to the output.
The discriminator is based on the self-attention [34], and the structure is shown in Fig 2. The mechanism can directly capture the correlation features between longer distance positions in the feature map output by the upper network, which is conducive to strengthening the model’s extraction of effective image information.
The specific network structure of the discriminative network is shown in Fig 3. The self-attention layer can obtain the output weights of features at different positions of the image through training, improve the computational efficiency of the discriminant network, reduce network stacking, and accelerate model convergence. In addition, in the actual training process of the network, the embedding layers do not interfere with the dimensionality of the input data, or the network structure.
We minimize the loss function for model convergence. The loss function of the model (both the generator and the discriminator) are as Eq (9). Where α denotes the low-count images and x denotes the high-count images.
In our implementation, we ran MLEM [35] for 25 to 35 iterations (up to general knowledge) and used its generative network output as the initial for α and x. The overall algorithm flowchart is presented in Algorithm 1. Where m and n denotes the number of max-iteration and sub-iteration, j means the image pixel.
Implement details
As the generative adversarial networks, it is inevitable that the update speed of two network parameters is inconsistent, which affects the convergence speed of the model. So, we use the TTUR (two-timescale update rule) [36] to enhance the training efficiency of the nets and that means the discriminator and the generator are given two different learning rates to balance the rate and achieve the final convergence. The learning rates are 4e-4 and 10e-4 separately for generator and discriminator by adjusting the update speed.
To avoid the over-fitting, we use dropout layers to the generator. Specifically, we trained the nets with 20% in the first three convolutional layers and the optimizer is Adam. All the networks are implemented in TensorFlow 1.0 and trained on NVIDIA GTX 1080Ti.
A. Setting of hyperparameter λ
The network was trained with λ set to 0.1,0.3,0.5,0.7 and 0.9. To reduce training time and computational costs, a randomly selected subset of 100 images was used for training and evaluation. The average sum of PSNR and SSIM was calculated to assess image quality, and the results are presented in Table 1.
According to Table 1, when λ is set to 0.5, the reconstructed image achieves the highest performance in both PSNR and SSIM metrics. However, as λ increases beyond this point, the quality of the reconstructed positron images begins to decline. Therefore, the hyperparameter λ in Eq (6) is ultimately set to 0.5.
B. Selection of training frequency
In the process of training, we can see in Fig 4(a) that the generative model tends to converge when trained about 300 epochs (each epoch includes 1000 steps). The training and validation mean squared errors (MSE) of the reconstruction network are shown in Fig 4(b), and we can see that the validation MSE almost to minimum at about 300 epochs and the iterative reconstruction model converges.
Experiment
Experimental data
In this study, the experimental data are obtained from GATE simulations. GATE is a Monte Carlo-based simulation software specifically designed for PET/SPECT applications and is one of the most widely used tools in the field of nuclear medicine imaging. Built on GEANT4 (General High Energy Physics Simulation Toolkit), GATE provides well-established high-energy physics models and comprehensive geometric modeling tools, enabling accurate replication of real PET imaging conditions.
In our experiment, we set the data of 10 seconds scan with 800 Bq dose. For industrial cavity defects that need to be detected, we set different templates, and partial examples are shown in Fig 5. Twelve of the templates for training the model as training samples, three templates as testing samples, and three templates as verification samples. Considering the characteristics of standardization of industrial parts, the templates here are set to more regular shapes. We also use the 30 seconds scan as the high-count data. At the same time, to enhance the data of the samples, we re-sample the high count and each file contains 1/4 of the number of the list-mode.
The system matrix is calculated by linear weighting [37] and the sampling data is reconstructed by MLEM algorithm. Here, the number of iterations depends on the prior knowledge of many experiments. Within the range of iterations, we can get the reconstructed positron image which is basically consistent with the actual description.
Experimental indicator
Here, to better describe the experimental results, we choose SSIM (structural similarity index) [38] and PSNR (Peak Signal to Noise Ratio) as the quantitative index to measure the quality of images. The indicators can be described as Eqs (10) and (11).
SSIM measures the similarity between two images and we can get a more intuitive comparison with structure, contrast and brightness. Where f(m, n) and represent the pixels of reconstructed images and real images respectively;
and
are the mean value of f(m, n) and
;
and
are the variance of f(m, n) and
;
is the covariance; c1 and c2 are the constant.
Where MAXI represents the maximum value of image color, and here the value is 255. mn is the size of the images and here the size is .
Experimental results
We compared the proposed method with some other methods, and the image reconstruction are based on MLEM algorithm. The specific description of the comparative model is as follows. Here, we use two indicators, PSNR and SSIM, to quantitatively evaluate the image reconstruction results of the model, and the numerical results obtained are shown in Table 2.
- MLEM: The current mainstream traditional algorithm for positron image reconstruction;
- MLEM+CNN: Combining the most commonly used convolutional neural network in deep learning field image processing with MLEM algorithm, using CNN to train prior knowledge of positron images;
- MLEM+GAN: Combining the original GAN with the MLEM algorithm, without improving the generation network and discriminant network, especially using the original adversarial loss function as a comparative model for ablation experiments.
- MLEM+SAGAN [39]: Combining the SAGAN model with the MLEM algorithm, unlike our method, SAGAN also introduces self-attention in the generator, this situation may make the network more conducive to capturing global features, but cannot achieve a balance between geometric and texture features, which is not conducive to the characterization of positron image features.
It can be seen from the data in Table 2 that the method proposed in this paper improves the two basic indexes of the image to a certain extent. Moreover, by analyzing the results of experiment, it is not difficult to find that in the process of simulation data, the more complex the image is, the more sampling information it contains, and the greater the quality improvement in the reconstructed image.
In addition, further analysis of the experimental results reveals that the more complex the mechanism for detecting industrial parts, the more feature information it contains. Here, three different types of templates were designed for comparison, as shown in Fig 6.
Among them, Derenzo phantom, which includes four different cylindrical shapes with different intervals and diameters (16mm, 13mm, 8mm, and 5mm). The circular ring template is a simple circular component with protrusions, where the outer and inner radii are 80mm and 75mm respectively, and the inner protrusion is 20mm long and 10mm wide, and the last template is a general complex industrial.
The reconstructed images obtained are shown in Fig 7, and the corresponding PSNR and SSIM results are shown in Table 3. Obviously, the algorithm proposed in this article improves the quality of more complex positron reconstruction images more significantly. The PIIR-GAN algorithm introduces a self-attention mechanism module in adversarial networks, which has a better capture effect on the global features of images. However, complex positron images clearly require higher global features, thus achieving a better improvement in reconstruction quality. Furthermore, as shown in Fig 7, the MLEM lacks deep feature extraction, resulting in limited reconstruction quality. MLEM+CNN improves local detail reconstruction, but suffers from insufficient global consistency. MLEM+GAN enhances visual realism but introduces artifacts. MLEM+SAGAN captures global features, but lacks fine structure. In contrast, the proposed PIIR-GAN algorithm obtains clearer reconstruction details with almost no blurring phenomenon, and has better high-frequency structural information.
To verify the training results and the practical reconstruction effect of our method in the field of industrial nondestructive testing, we design a group of experiments based on the industrial parts of hydraulic cylinder. Hydraulic cylinder can convert hydraulic energy into mechanical energy, which is widely used in various mechanical hydraulic systems. Therefore, it has good application value to carry out the actual test. In the experiment, the PET detector we used was Trans-PET Explorist 180, with a detector crystal resolution of 1 millimeter. The hydraulic component is made of aluminum, with an outer diameter of 55mm, an inner diameter of 45mm, a wall thickness of 5mm, and an axial length of 20cm. We cut a groove with a depth of 3mm at the pipe wall of the hydraulic component to simulate possible cracks. During the experiment, approximately 350 milliliters of radioactive mixture with a configuration of 1.85 mCi were injected. The reconstruction results under different comparative models are shown in Fig 8.
From the local enlarged drawing, we can see the crack shape of the inner wall of the hydraulic parts. Combined with the specific size of the hydraulic parts (Length, diameter, etc.), we can further determine the location and size of the crack. Industrial parts generally have standard size parameters, so the operability of this method is extremely high. Here, it also needs to be explained, due to the limitations of current hardware conditions, the resolution of the image needs to be improved. Of course, in the later researches, we can also strengthen the research on the post-processing stage of positron reconstruction image, especially in the aspect of edge processing, so as to detect the position of defects more accurately in the post-processing stage of image, and better achieve the purpose of industrial nondestructive testing.
Discussion
At present, our application of PET technology in the field of industrial non-destructive testing is mainly focused on the gaps in complex cavities and the description of the internal flow field of industrial parts. Many existing works have used pure MLEM algorithm, simple convolutional neural network or other ways to obtain better images. Here, we use Generative Adversarial Networks as the sampling data representation and the model is added in the iterative reconstruction stage. Thus, compared with other methods, especially in the post-processing stage, the proposed method is constrained in the data sampling which can retain finer information features to avoid the loss of details in the reconstruction process. At the same time, in the iterative reconstruction process, it can be found that as the number of iterations increases, the image quality will not be significantly improved, and may even lead to an increase in noise. Therefore, we ensure the stability of the model through regular optimization of the network.
In addition, we observed that the adversarial learning mechanism in PIIR-GAN effectively improves the model’s ability to distinguish signal from noise, leading to a clearer reconstruction of structural boundaries. This advantage becomes more pronounced under low-sampling conditions, where traditional iterative or CNN-based methods tend to produce blurred edges or over-smoothed textures. The integration of prior information through the generator ensures that the reconstructed images preserve both global consistency and local details, which is essential for accurately identifying defect morphology in industrial components.
In addition, we designed a more targeted experiment to prove the superiority of the proposed method in image detail feature processing. Two different sets of models were designed using SolidWorks in the experiment to simulate the possible presence of foreign objects of different shapes in the inner cavity of a circular pipeline. The models are shown as Fig 9.
And the comparison model in the experiment is the same as above, and the final positron image obtained is shown in Fig 10.
The above two groups of positron images are obtained in the practical application. Specifically, the same shape of the cavity is detected. From the experimental results, we can see that our method can describe the wire more clearly and completely, and there is no or very little image layer breaking phenomenon in the middle, and the experimental effect is not affected by the different shapes of the wire. And the quantitative comparison is shown in Table 4. Clearly, PIIR-GAN has shown significant improvements in PSNR and SSIM metrics and have also shown good performance improvements in practical industrial detecting.Specifically, PIIR-GAN achieved an average PSNR improvement of 3.3dB and a 6% increase in SSIM compared with the baseline GAN model, confirming its enhanced image fidelity and structural preservation. Overall, these results validate that incorporating adversarial learning into the iterative PET reconstruction framework provides both quantitative and qualitative benefits, making it more suitable for industrial non-destructive testing scenarios.
Conclusions
Considering the objective difficulties faced by positron imaging technology in industrial non-destructive testing applications, such as short adoption time and limited data, to solve the problem of poor image reconstruction quality, this paper proposes an industrial positron image iterative reconstruction algorithm that integrates generative adversarial networks. This algorithm can significantly improve the quality of positron image reconstruction and has strong domain applicability. PIIR-GAN showed that combined with the deep neural network, adding prior knowledge to MLEM reconstructing algorithm can better adapt to the specific conditions on industrial non-destructive testing. In particular, we introduced a self-attention mechanism in the network to make PET image reconstruction in line with actual image characteristics (A large number of experiments show that the same set of PET detection device, that is, under the same physical parameters, the network structure is basically fixed and not affected by the images).
However, the proposed approach still has some limitations that would need to be addressed. First, considering the short sampling time and the difficulty in data collection, the network model needs to be further improved. Second, most experiments are under simulation conditions, and more field experiments are needed to optimize the entire reconstruction algorithm network. In addition, on the basis of optimizing the image reconstruction algorithm, we also need to consider the use of denoising, edge detection and other methods in the subsequent processing stage of the image to improve the final presentation quality of the image, so as to achieve more accurate detection. In the future, more work will focus on improving the quality of reconstructed images. We plan to study more advanced neural network models and further perfect the images dataset. Moreover, the three-dimensional reconstruction of images is also the focus of the subsequent research.
Supporting information
S1 Data. We have uploaded the data set as a Supporting information file named “data.zip”.
https://doi.org/10.1371/journal.pone.0335912.s001
(ZIP)
References
- 1. Chen S, Liu H, Shi P, Chen Y. Sparse representation and dictionary learning penalized image reconstruction for positron emission tomography. Phys Med Biol. 2015;60(2):807–23. pmid:25565039
- 2. Tahaei MS, Reader AJ. Patch-based image reconstruction for PET using prior-image derived dictionaries. Phys Med Biol. 2016;61(18):6833–55. pmid:27581747
- 3. Tang J, Rahmim A. Bayesian PET image reconstruction incorporating anato-functional joint entropy. Phys Med Biol. 2009;54(23):7063–75. pmid:19904028
- 4.
Nguyen V-G, Lee S-J. Anatomy-based PET image reconstruction using nonlocal regularization. In: SPIE Proceedings. 2012. 83133T. https://doi.org/10.1117/12.911690
- 5. Wang G, Qi J. PET image reconstruction using kernel method. IEEE Trans Med Imaging. 2015;34(1):61–71. pmid:25095249
- 6. Hutchcroft W, Wang G, Chen KT, Catana C, Qi J. Anatomically-aided PET reconstruction using the kernel method. Phys Med Biol. 2016;61(18):6668–83. pmid:27541810
- 7. Novosad P, Reader AJ. MR-guided dynamic PET reconstruction with the kernel method and spectral temporal basis functions. Phys Med Biol. 2016;61(12):4624–44. pmid:27227517
- 8. Ellis S, Reader AJ. Simultaneous maximum a posteriori longitudinal PET image reconstruction. Phys Med Biol. 2017;62(17):6963–79. pmid:28643694
- 9. Goodfellow I, Pouget-Abadie J, Mirza M,Xu B,Warde-Farley D, et al. Generative adversarial nets. Advances in neural information processing systems. 2014; 2672–80.
- 10. Kang E, Min J, Ye JC. A deep convolutional neural network using directional wavelets for low-dose X-ray CT reconstruction. Med Phys. 2017;44(10):e360–75. pmid:29027238
- 11. Kang E, Chang W, Yoo J, Ye JC. Deep convolutional framelet denosing for low-dose CT via wavelet residual network. IEEE Trans Med Imaging. 2018;37(6):1358–69. pmid:29870365
- 12. Chen H, Zhang Y, Kalra MK, Lin F, Chen Y, Liao P, et al. Low-dose CT with a residual encoder-decoder convolutional neural network. IEEE Trans Med Imaging. 2017;36(12):2524–35. pmid:28622671
- 13. Zhang J, Cui Z, Jiang C. Hierarchical organ-aware total-body standard-dose PET reconstruction from low-dose PET and CT images. IEEE Transactions on Neural Networks and Learning Systems. 2013.
- 14. Ote K, Hashimoto F, Onishi Y, Isobe T, Ouchi Y. List-mode PET image reconstruction using deep image prior. IEEE Trans Med Imaging. 2023;42(6):1822–34. pmid:37022039
- 15.
Sun Y, Mawlawi O. Legopet: hierarchical feature guided conditional diffusion for pet image reconstruction. In: 2025 IEEE 22nd International Symposium on Biomedical Imaging (ISBI). 2025. p. 1–5. https://doi.org/10.1109/isbi60581.2025.10980656
- 16.
Huang B, He B, Chen Y, et al. Diffusion transformer meets random masks: an advanced PET reconstruction framework. In: arXiv preprint 2025. https://doi.org/arXiv:2503.08339
- 17. Chen X, Duan Y, Houthooft R. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. Advances in Neural Information Processing Systems. 2016;2016:2172–80.
- 18.
Wang X, Yu K, Wu S, et al. Esrgan: Enhanced super-resolution generative adversarial networks. Proceedings of the European Conference on Computer Vision (ECCV). 2018.
- 19.
Ledig C, Theis L, Huszár F, et al. Photo-realistic single image super-resolution using a generative adversarial network.Proceedings of the IEEE conference on computer vision and pattern recognition. 2017; p. 4681–90.
- 20. Jolicoeur-Martineau A. The relativistic discriminator: a key element missing from standard GAN. arXiv preprint 2018. arXiv:1807.00734
- 21. Nie D, Trullo R, Lian J, Petitjean C, Ruan S, Wang Q, et al. Medical image synthesis with context-aware generative adversarial networks. Med Image Comput Comput Assist Interv. 2017;10435:417–25. pmid:30009283
- 22. Guibas JT, Virdi TS, Li PS. Synthetic medical images from dual generative adversarial networks. arXiv preprint 2017. https://arxiv.org/abs/1709.01872
- 23.
Nie D, Trullo R, Lian J. Learning myelin content in multiple sclerosis from multimodal MRI through adversarial training. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. 2018. p. 514–22.
- 24.
Olut S, Sahin YH, Demir U, et al. Generative adversarial training for MRA image synthesis using multi-contrast MRI. International workshop on predictive intelligence in medicine. Cham: Springer; 2018. p. 147–54.
- 25.
Isola P, Zhu JY, Zhou T. Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017. p. 1125–34.
- 26.
Zhao Z, Li B, Dong R. A surface defect detection method based on positive samples. In: Pacific Rim International Conference on Artificial Intelligence. 2018. p. 473–81.
- 27. Xue Y, Bi L, Peng Y, Fulham M, Feng DD, Kim J. PET synthesis via self-supervised adaptive residual estimation generative adversarial network. IEEE Trans Radiat Plasma Med Sci. 2024;8(4):426–38.
- 28.
Cui J, Wang Y, Wen L. Image2points: a 3d point-based context clusters gan for high-quality pet image reconstruction. In: ICASSP 2024 -2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2024. p. 1726–30.
- 29. Zheng Q, Zhao P, Li Y, Wang H, Yang Y. Spectrum interference-based two-level data augmentation method in deep learning for automatic modulation classification. Neural Comput & Applic. 2020;33(13):7723–45.
- 30. Zheng Q, Zhao P, Zhang D, Wang H. MR-DCAE: Manifold regularization-based deep convolutional autoencoder for unauthorized broadcasting identification. Int J of Intelligent Sys. 2021;36(12):7204–38.
- 31. Zheng Q, Tian X, Yang M, Wu Y, Su H. PAC-Bayesian framework based drop-path method for 2D discriminative convolutional network pruning. Multidim Syst Sign Process. 2019;31(3):793–827.
- 32.
Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. 2015. p. 234–41.
- 33. Oktay O, Schlemper J, Folgoc LL, et al. Attention U-Net: learning where to look for the pancreas. arXiv preprint 2018. arXiv:1804.03999
- 34.
Wang X, Girshick R, Gupta A, He K. Non-local neural networks. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2018. p. 7794–803. https://doi.org/10.1109/cvpr.2018.00813
- 35. Shepp LA, Vardi Y. Maximum likelihood reconstruction for emission tomography. IEEE Trans Med Imaging. 1982;1(2):113–22. pmid:18238264
- 36.
Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Conference and Workshop on Neural Information Processing Systems. 2017. p. 6629–40.
- 37. Wang H, Kwong S, Kok C-W. An efficient mode decision algorithm for H.264/AVC encoding optimization. IEEE Trans Multimedia. 2007;9(4):882–8.
- 38. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process. 2004;13(4):600–12. pmid:15376593
- 39.
Zhang H, Goodfellow I, Metaxas D, et al. Self-attention generative adversarial networks. International conference on machine learning. PMLR; 2019. p. 7354–63.