Figures
Abstract
PET is a popular medical imaging modality for various clinical applications, including diagnosis and image-guided radiation therapy. The low-dose PET (LDPET) at a minimized radiation dosage is highly desirable in clinic since PET imaging involves ionizing radiation, and raises concerns about the risk of radiation exposure. However, the reduced dose of radioactive tracers could impact the image quality and clinical diagnosis. In this paper, a supervised deep learning approach with a generative adversarial network (GAN) and the cycle-consistency loss, Wasserstein distance loss, and an additional supervised learning loss, named as S-CycleGAN, is proposed to establish a non-linear end-to-end mapping model, and used to recover LDPET brain images. The proposed model, and two recently-published deep learning methods (RED-CNN and 3D-cGAN) were applied to 10% and 30% dose of 10 testing datasets, and a series of simulation datasets embedded lesions with different activities, sizes, and shapes. Besides vision comparisons, six measures including the NRMSE, SSIM, PSNR, LPIPS, SUVmax and SUVmean were evaluated for 10 testing datasets and 45 simulated datasets. Our S-CycleGAN approach had comparable SSIM and PSNR, slightly higher noise but a better perception score and preserving image details, much better SUVmean and SUVmax, as compared to RED-CNN and 3D-cGAN. Quantitative and qualitative evaluations indicate the proposed approach is accurate, efficient and robust as compared to other state-of-the-art deep learning methods.
Citation: Zhao K, Zhou L, Gao S, Wang X, Wang Y, Zhao X, et al. (2020) Study of low-dose PET image recovery using supervised learning with CycleGAN. PLoS ONE 15(9): e0238455. https://doi.org/10.1371/journal.pone.0238455
Editor: Mathieu Hatt, INSERM, FRANCE
Received: January 3, 2020; Accepted: August 17, 2020; Published: September 4, 2020
Copyright: © 2020 Zhao et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data are available within the paper and on Figshare: https://figshare.com/articles/dataset/LDPET/12686099.
Funding: MinFound Medical Systems Co., Ltd provided support for this study in the form of salaries for authors HY, LZ, SG, XW, and YW. The specific roles of these authors are articulated in the ‘author contributions’ section. The funders played a role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. No additional external funding was received for this study.
Competing interests: The authors have read the journal’s policy and have the following competing interests: HY, LZ, SG, XW, and YW are paid employees of MinFound Medical Systems Co., Ltd. There are no patents, products in development or marketed products associated with this research to declare. This does not alter our adherence to PLOS ONE policies on sharing data and materials.
Introduction
Positron Emission Tomography (PET) is a widely used imaging modality for various clinical applications, such as lesion malignancy, disease stage, and treatment monitoring [1–3]. Compared with computed tomography (CT) and magnetic resonance imaging (MRI), PET is a functional imaging technique that detects the metabolism processes of human body [4]. To reach a certain PET image quality for diagnostic purposes, a typical dose of injected radioactive tracers usually ranges from 185∼555 MBq, depending on PET scanners, protocols, reconstruction methods, patients and so on. Since high gamma radiation dosage in a patient may induce genetic damages and cancerous diseases [5–7], it inevitably raises concerns about the potential higher risk of radiation exposure damage. Thus, it is desirable to reduce the dose of radioactive tracers in PET imaging. However, the major drawback of dose reduction is that higher noise, worse contrast and information loss may be involved in the reconstructed images, resulting in an inferior image quality and unreliable diagnosis.
A series of methods has been proposed to improve the image quality for the low-dose PET (LDPET) imaging, while preserving crucial diagnosis information. Those algorithms can be roughly categorized into traditional methods such as iterative reconstruction algorithms [8, 9], post-processing methods [10–13] and deep learning based methods [14–22]. In general, those strategies for improving PET image quality are either hardware-oriented or computationally intensive. Besides, the LDPET image contains more complex spatial variations, correlations and statistical noise than the full-dose PET (FDPET) image, which limits the performance of the traditional methods.
Recently, deep learning has drawn a mount of attention in computer vision applications and medical image analysis areas [4–7, 23]. For instance, the image classification [24] and face verification [25] can achieve human-level performance. Algorithms based on deep learning have made some success in low-dose CT (LDCT) reconstruction and denoising [14–18]. These methods learn a non-linear mapping from a LDCT image to high-quality CT image to recover missing high-frequency details. While in recovering or denoising LDPET, there are much fewer works with deep learning methods reported. Xiang et al. [19] proposed a deep auto-context CNN model that synthesized a high quality image from 1/4 of FDPET image and corresponding MR T1-image. Xu et al. [20] used a U-Net [26] like network to recover a full-dose quality PET image from 1/200 of FDPET image, and applied a multi-slice input strategy to make the network more robust to noise. Wang et al. [21] designed an end-to-end framework based on 3D conditional GANs (3D-cGANs) to estimate the high-quality PET image from the corresponding LDPET image. The 3D convolution operation makes the model avoid the discontinuous cross-like artifacts that usually occurs in 2D convolution based models. Kaplan et al. [22] proposed a deep learning model that takes specific image features into account in the loss function to denoise 1/10 of FDPET image. Chen et al. [27] proposed to combine both PET and MR information to synthesize high quality and accurate PET images. More recent work from Ouyang et al. [28] suggests that combining a generative adversarial network (GAN) with feature matching into the discriminator can lead to similar performance even without the MR information.
Rather than using deep learning method as a post-processing tool, Gong et al. [29] proposed a residual convolutional auto-encoder within a Machine Learning framework to denoise PET images. More recently, Haggstrom et al. [30] took the PET sinogram data as the input and directly generate PET reconstructed images, highlighting a 100-fold speedup for reconstruction compared to standard iterative techniques such as ordered subset expectation maximization (OSEM).
In general, physicians use both the maximum SUV (SUVmax) and the mean SUV (SUVmean) to characterize the high uptakes [31], but SUVmax is more often used in practice since SUVmean heavily depends on volume of interest (VOI) selected, while SUVmax value is unique and reproducible in VOI [32, 33]. Inspired by the most recent advanced neural networks, such as DenseNet [34], Residual CNN [35], and CycleGAN [36], a cycle Wasserstein regression adversarial training framework, named S-CycleGAN, is proposed and studied for the PET brain imaging in this paper. Although some good performance in recovering or denoising LDPET images were reported, those deep learning based methods mentioned above were not evaluated quantitatively for lesion SUVs, which limited their usage in clinical applications. In order to evaluate the clinical performance of our model, we also proposed a simulation framework to produce a series of simulation data to mimic complex clinical situations. The S-CycleGAN model was then applied to the clinical and simulated LDPET datasets (10% and 30% of FDPET datasets), and studied qualitatively and quantitatively.
Methods
The goal of this work is to train a model to learn the non-linear mapping between LDPET and FDPET images. As shown in Fig 1, the proposed network is based on a CycleGAN architecture.
Overview of the proposed framework for LDPET image recovery.
The proposed model includes two generators and two discriminators. We denote GAB is the mapping from LDPET domain (A) to FDPET domain (B), the GBA represents the opposite direction mapping. In addition, there are two discriminators DA and DB which intend to identify whether the output of each generator is real or fake. Then, we train the generators and discriminators simultaneously. Thus, we have the following optimization problem:
(1)
Our proposed network combines four types of loss functions: adversarial loss (), cycle-consistency loss (
), identity loss (
) and supervised learning loss (
). Therefore, the overall loss is defined by:
(2)
where α, β and γ are hyperparameters.
Adversarial loss: We employ adversarial losses to generate image samples to obey the empirical distributions in the source and target domains. To improve the training stability of GANs, we apply the 1-Wasserstein distance [37] instead of the original log-likelihood function. The 1-Wasserstein distance or Earth-Mover (EM) distance is defined as follows.
(3)
Where Π(Pr, Pg) denotes the set of all joint distributions γ(x, y) whose marginals are respectively Pr and Pg.
Thus, the adversarial objective function is defined as follows.
(4)
Where λ is a regularization parameter, which controls the trade-off between the Wasserstein distance and the gradient penalty term. is uniformly sampled along straight lines for pairs of GAB(xA) and xB. The adversarial loss for the reverse direction
is defined in a similar way. The final adversarial loss(
) is defined
(5)
Cycle consistency loss: We adopt a cycle consistency term that the FDPET and LDPET images could be transformed mutually as an additional regularization to help learning of GAB and GBA. The cyclic loss is defined by
(6)
Where ∥:∥1 denotes the l1-norm. This allows for additional information to be shared between LDPET and FDPET images in learning their corresponding generators.
Identity loss: In real clinical situation, the input to the generator GAB can be a full-dose image, but we expect the generator does not alter such clean image, and vice versa. Besides, indentity loss provides another regularization in the training procedure and is formulated as follows:
(7)
Supervised learning loss: Since we have paired datasets, we can train our model in a supervised fashion. Then, we can define a supervision loss as follows:
(8)
Network architecture
Our proposed model, S-CycleGAN, is constituted of two generator networks, GAB and GBA, and two discriminator networks, DA and DB. The generator networks take one domain’s image and estimate another domain’s image. The discriminator networks aim to differentiate between the real and estimated image.
Generative networks: The network architecture of two generators GAB and GBA is illustrated in Fig 2. The basic structure is optimized for the LDCT image denoising in [38, 39]. To reduce network complexity and adapt to PET image, we set the filter number to 64 instead of 128 in the original model and add a ReLU layer before model output. As shown in Fig 2, the first two convolution layers use 64×3×3 convolution kernels to produce 64 feature maps, and connect to 6 sets of residual modules, where each module is composed of 3 sets of convolution, batch normalization, and a ReLU layer, and one residual connection with a ReLU layer. Later on, a concatenation layer that concatenates the inputs of each module and the output of the last module, and two convolution layer with 64 feature maps are applied. Finally, the last convolution layer with 3×3 convolution kernel combined with an end-to-end bypass connection and an additional ReLU layer are used to estimate the FDPET image.
A generator architecture in the proposed framework.
Discriminator: The discriminators take either a real PET image or an estimated one as input, and determines whether the input is real or not. As shown in Fig 3, the discriminator network is designed to have 4 stages of convolutions, followed by two fully-connected layers, of which the first has 1024 outputs and the last has 1 output. We apply 4×4 filter size for the convolution layers which have different numbers of filters as 64, 128, 256, 512 respectively. In addition, we use Leaky ReLU activation in the discriminator for all layers, with slope 0.2.
A discriminator architecture in the proposed framework.
Experimental setup
Datasets
We trained our model by using human brain datasets. A total of 109 clinic patient (range 44.3-103kg) PET/CT images were taken by the Minfound ScintCare PET/CT 720L scanner with injection of 370.81±64.38 MBq of 18F-fluorodeoxyglucose(FDG), and we randomly selected 89, 10 and 10 patient data for training, validation and testing, respectively. All scans were taken about 5 minutes and usually start at 45-60 minutes later after injection. The reconstruction was performed using the manufacturer-provided software with all physical corrections, including attenuation, scatter, randoms, dead-time and SUV correction. The size of each 3D reconstructed PET image is 192×192×96 with pixel size of 2.1 mm. FDPET and LDPET images were reconstructed with the same parameters and post filters to ensure comparable spatial resolution in both images. Two different simulated doses, i.e. 10% and 30% counts of original scan, where generated by randomly discarding events in FDPET list mode data. Although the 10% and 30% images were generated by emulated low-count scans, however, those images have comparable quality with actual low-dose scan, and confirmed by the recent work [40]. By this way, FDPET and LDPET images are spatially aligned.
In order to evaluate the clinical feasibility of our proposed model, a Monte Carlo simulation framework using GATE [41, 42] was carefully designed and shown in Fig 4. At the first step, lesion maps with different shapes, sizes, and locations were extracted from a few of known patient’s datasets (different from the above 109 patients), and a patient’s attenuation map (μ-map) was generated from its corresponding CT image. Then, these two maps were fed into GATE, and were simulated with the same system settings as Minfound ScientCare PET/CT 720L scanner. Finally, the simulated coincidence data of lesions combined with the clinical coincidence data of the patient were reconstructed by the manufacturer-provided software to produce the final PET image. To systematically evaluate model performance, a series of simulation data with various activities, sizes and shapes were produced and reconstructed. In order to reduce statistical variations, each simulation configuration was repeated 3 times and total 45 simulations were used in later quantitative evaluations. The details of those lesions are also provided in Table 1 and Fig 5. The activity for background is about 30 kBq/ml. Some typical simulated PET images are shown in Fig 6.
A GATE based simulation architecture.
The distribution of 5 embedded lesions in the GATE simulation.
Typical FDPET images with embedded lesions using the proposed simulation architecture.
Where Dmax and Dmean are the maximum and mean diameters, respectively.
In order to reduce the computational cost for training, we extracted overlapping patches from LDPET and FDPET images instead of directly feeding entire PET images to the training pipeline. We cropped LDPET and FDPET images into patches of 56×56 at the same place for the supervised learning with sliding step 40. In total, there are 136704 and 15360 patches for training and validation. Since PET images have a large range in pixel values, we scaled pixel values to [0, 1].
Evaluation measures
Six measures are used to evaluate the model performance including the normalized root mean square error (NRMSE), structural similarity index (SSIM [43]), peak signal-to-noise ratio (PSNR), learned perceptual image patch similarity (LPIPS) [44], relative errors (RE) for SUVmean and SUVmax, which are defined as following.
(9)
(10)
(11)
(12)
Where C1 and C2 are constants, μx, μy, σx, σy, and σxy are mean and standard deviation in the patch centered at pixel (i,j). MAX is the peak intensity of the image, MSE is absolute mean square error.
The SUV (standardized uptake value) is commonly used as a relative measure of FDG uptake. The basic expression for SUV [32] is
(13)
Where r is the radioactivity activity concentration [kBq/ml] measured by the PET scanner within a region of interest (ROI), a′ is the decay-corrected amount of injected radiolabeled FDG (kBq), and w is the weight of the patient (g), which is used as a surrogate for a distribution volume of tracer.
Implementation details
In the proposed model, training was performed by minimizing the loss function 2. We utilized the Adam optimizer [45] with β1 = 0.5 and β2 = 0.999 to minimize the total loss function of the proposed network. We set the learning rate to 2×10−4, hyperparameters α = 10, β = 5, and γ = 5. The trade-off parameter λ between Wasserstein distance and gradient penalty was set to be 10, as [37] suggested. The hyperparameters basically were derived from the original CycleGAN paper. As for the parameter, γ, it was determined by experiment in order to get a trade-off between the noise and the bias of SUV at the lesion regions. The size of the patch was set to 56×56 and the mini-batch is 16. Kernels were initialized randomly from a Gaussian distribution. All experiments were conducted using Keras [46] with Tensorflow backend on a NVIDA TITAN GTX GPU.
The training epoch was set to 200 based on experience with early-stop strategy when the validation loss is minimal (the patience value is 5). It takes 7 days for training at current GPU hardware. Although the training was done on patches, the proposed network can process images of arbitrary sizes. All the testing images were simply fed into the network without decomposition and required 74ms of inference time per image slice.
Experimental results
Comparison with other methods
To study the effectiveness of our proposed model, we compared it with RED-CNN [15] and 3D-cGAN [21]. The network structure and parameters of these competing methods were set per the suggestions from the original papers and re-implemented by Keras. For a qualitative comparison, some sample images of the predicted FDPET from three deep learning methods, the corresponding LDPET and FDPET reconstruction are shown in Figs 7 and 8 for 10% and 30% dose of FDPET, respectively. The estimated images by all deep learning methods show better image quality than low-dose images, providing better noise reduction and structure details recovery.
Qualitative comparison of PET images by S-CycleGAN, RED-CNN and 3D-cGAN, where low-dose data is 10% of FDPET.
Qualitative comparison of PET images by S-CycleGAN, RED-CNN and 3D-cGAN, where low-dose data is 30% of FDPET.
The quantitative measures in terms of NRMSE, SSIM and PSNR are shown in Table 2 using 10 testing patient datasets. All three predicted images have better noise control and structure similarities than low-dose images, but similar peak signal to noise ratios. RED-CNN and 3D-cGAN models have better NRMSE scores than S-CycleGAN, however, their predicted images suffer from over-smoothing issues and may compromise the diagnostic performance, as shown in Figs 8 and 7 (indicated by red arrows).
As suggested by Zhang et al. [47], the traditional metrics (L2/PSNR, SSIM, FSIM) disagree with human judgments, a learned perceptual image patch similarity metric was proposed to evaluate image quality. The LPIPS measurements between model prediction and FDPET is shown in Fig 9. The estimated images by all deep learning methods show better LPIPS scores than low-dose images and S-CycleGAN obtains the best score. The average LPIPS scores of LDPET (30% of FDPET), S-CycleGAN, RED-CNN and 3D-cGAN are 0.035, 0.026, 0.031 and 0.031, respectively.
LPIPS score (smaller is better) comparison of PET images by LDPET (30% of FDPET), S-CycleGAN, RED-CNN and 3D-cGAN.
Clinical evaluation for specific VOIs
In clinic, the mean and maximum SUVs are often used as bases for diagnosis to characterize suspicious high uptakes [31, 32]. Therefore, the SUV measures are used to investigate the effectiveness of the proposed method for specific VOIs on both the normal and lesion tissues. The datasets were produced by our proposed simulation framework as mentioned in Experimental Setup section. In this analysis, the mean and maximum SUV biases and deviations were evaluated for all deep learning models mentioned above.
The average biases and standard deviations of SUVmean and SUVmax of lesion tissues are shown in Tables 3, 4 and Tables 5, 6, respectively. Since SUVmax is not critical for normal tissues, only SUVmean error is shown in Table 7. The results of different lesion sizes and FDG concentrations are also shown in above tables.
SUVmean deviation: As observed in Table 7, all the models have very similar SUV-mean values in normal tissues which biases are less than 5% for both 10% and 30% dose levels. However, as seen in Tables 3 and 4, the RED-CNN and 3D-cGAN have much larger biases than S-CycleGAN in lesion tissues, especially for smaller lesion sizes and lower activities. The average SUVmean biases of S-CycleGAN, RED-CNN and 3D-cGAN for all lesions and activities are -6.4±5.3%, -18.7±11.8% and -20.0±10.8% for 10% dose level, and -2.8±4.1%, -6.3±6.4% and -9.8±6.0% for 30% dose level, respectively. It can be also seen that the biases and deviations of SUVmean decrease as the lesion size and activities increases for S-CycleGAN model in most cases. Those observations indicate the good robustness of our proposed model.
SUVmax deviation: The SUVmax results of all three deep learning methods are shown in Tables 5 and 6 for 10% and 30% dose of FDPET, respectively. Since the single pixel value in the VOI is largely affected by the statistical property of data, the SUVmax values in LDPET images have large biases and deviations, especially for the lower dose level. Our proposed S-CycleGAN model is trending to reduce the biases and deviations but this ability gets worse as lesion sizes decrease. The average SUVmax biases of S-CycleGAN, RED-CNN and 3D-cGAN for all lesions and activities are -3.7±16.2%, -24.9±11.7% and -28.0±16.5% for 10% dose level, and -5.2±6.8%, -11.4±11.7% and -14.8±9.8% for 30% dose level, respectively. Those results are suggesting the S-CycleGAN method can better preserve the SUVmax values than other two methods.
Ablation study
Impact of supervised learning loss: The impact of the supervised learning loss was studied for the proposed model. A modified model, named as CycleGAN, was trained and tested with all the loss functions except the supervised loss Lsup. Image artifacts of missing structures are observed in about 7% of the slices generated by the CycleGAN model, as indicated by the red and yellow rectangle in Fig 10. Therefore, the use of supervised learning loss could reduce these artifacts and maintaining the fidelity of the PET image.
A typical PET slice image with artifact from CycleGAN.
Impact of cycle-consistency loss: The effectiveness of cycle consistence loss was also studied by comparing the S-CycleGAN and 3D-cGAN model which didn’t involve this loss. As shown in Tables 3, 4, 5 and 6, S-CycleGAN model can better preserve the SUVmean and SUVmax values than 3D-cGAN, which indicates the effectiveness and necessity of the cycle consistence loss even if it is originally designed for the unpaired datasets training.
Discussion
In order to systematically evaluate model performance, we designed a novel simulation framework to produce clinical-like data, in which the embedded lesions are exacted from the clinical data by considering realistic structures, sizes, activities and dose levels. Such simulations would be helpful to understand the clinical performance of the proposed method since it is almost impossible to know the true lesion uptakes in clinic. Moreover, this method can be extend to other related model performance studies.
Although our model has achieved the compelling results, there still exist some limitations. Our proposed model, S-CycleGAN, requires longer training time than other standard GAN-based and CNN-based methods. The future work should consider more efficient architectures. Though this paper mainly focuses on PET brain images, the same model with different hyperparameters has been applied to PET body images too. More results will be presented in the near future once enough PET body datasets are acquired and trained.
Recent published papers [20–22] on LDPET image recovery were extended to even lower doses. However, it is difficult to conclude which approach can reduce more dose since different paper uses different datasets, acquisition protocols and scanners. In this paper, the training set of 99 patients has 110±23M average coincidence counts. Consequently, our proposed S-CycleGAN model actually takes the count variation into account in the training and can be used for a relatively widespread dose levels in complicated clinical situations. The recently published paper [48] uses a very similar method, the CycleGAN, for LDPET denoising, but they still didn’t investigate the metrics of SUVmax and robustness to different count levels.
All of those approaches usually compared the structure similarity, noise, signal-to-noise ratio or SUVmean, but none of them involved the evaluation on SUVmax. SUVmax is more often used in the clinical practice due to its better reproducibility than SUVmean since the maximum value within a VOI (or region-of interest) is invariant with respect to small spatial shifts [32, 33]. Due to the supervised training mode, the SUVmean can be easily preserved, but not the SUVmax. From systematic study of SUVmean and SUVmax, our proposed model has demonstrated promising results in recovering a high quality image from a LDPET image. However, smaller lesion size and lower activity actually degrade the performance of all models compared in this paper.
As shown in Tables 3, 4, 5 and 6, the SUVmean values can be relatively easier to be preserved as compared the SUVmax values. Our proposed model has demonstrated better quantitative results than RED-CNN and 3D-cGAN no matter which dose level is used. When predicted images in the same dose level are compared, the SUVmean values show strong dependence on lesion sizes and activity concentrations. On the other hand, the SUVmax values only show strong dependence on lesion sizes, but not the activity concentrations. Moreover, the SUVmax values still have quite large variations even though 45 simulations are used in evaluations. These phenomena can be partially explained by two factors: one is image noise caused by data itself and reconstruction/post-processing methods, which can strongly affect SUVmax values relying only on the single pixel values; another is the partial volume effect caused by the finite system spatial resolution and image sampling, which can heavily reduce the accuracy of SUVmean and SUVmax values, especially for the smaller volumes of VOIs, or the lower activity ratio between VOIs and their surrounding background [49].
As compared to the 30% dose level, the images recovered from the 10% dose level still have good scores in normal tissues in terms of NRMSE, SSIM and PSNR, but much larger biases and deviations for the SUVmean and SUVmax in lesion tissues for all three deep learning methods. This alerts us a potential risk that any diagnosis relying on these two indexes could be changed in clinical practice. Therefore, we should be cautious in developing any deep learning approaches which could largely change SUVmean and SUVmax while reducing dose. For this reason, the 30% dose level is preferred in this study since it can better balance the tradeoff between SUV values and dose reduction.
Conclusion
In conclusion, we have introduced a novel deep learning based generative adversarial model with the cycle consistent to estimate the high-quality image from the LDPET image. The proposed S-CycleGAN approach has produced comparable image quality as corresponding FDPET images by suppressing image noise and preserving structure details in a supervised learning fashion. Systemic evaluations further confirm that the S-CycleGAN approach can better preserve mean and maximum SUV values than other two deep learning methods, and suggests the amount of dose reduction should be carefully decided according to the acquisition protocols and clinical usages.
Acknowledgments
The authors would like to thank many colleagues at Zhejiang University and MinFound Medical Systems Co., Ltd. for the helpful discussion and PET image reconstruction.
References
- 1. Juweid M. Imaging Subcommittee of International Harmonization Project in Lymphoma. Use of positron emission tomography for response assessment of lymphoma: consensus of the Imaging Subcommittee of International Harmonization Project in Lymphoma. J Clin Oncol. 2007;25:571–578. pmid:17242397
- 2. Avril NE, Weber WA. Monitoring response to treatment in patients utilizing PET. Radiologic Clinics. 2005;43(1):189–204. pmid:15693656
- 3. Fletcher JW, Djulbegovic B, Soares HP, Siegel BA, Lowe VJ, Lyman GH, et al. Recommendations on the use of 18F-FDG PET in oncology. Journal of Nuclear Medicine. 2008;49(3):480–508. pmid:18287273
- 4. Kamnitsas K, Ledig C, Newcombe VF, Simpson JP, Kane AD, Menon DK, et al. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Medical image analysis. 2017;36:61–78. pmid:27865153
- 5. Kleesiek J, Urban G, Hubert A, Schwarz D, Maier-Hein K, Bendszus M, et al. Deep MRI brain extraction: a 3D convolutional neural network for skull stripping. NeuroImage. 2016;129:460–469. pmid:26808333
- 6. Valverde S, Cabezas M, Roura E, González-Villà S, Pareto D, Vilanova JC, et al. Improving automated multiple sclerosis lesion segmentation with a cascaded 3D convolutional neural network approach. NeuroImage. 2017;155:159–168. pmid:28435096
- 7. Dolz J, Desrosiers C, Ayed IB. 3D fully convolutional networks for subcortical segmentation in MRI: A large-scale study. NeuroImage. 2018;170:456–470. pmid:28450139
- 8. Shepp LA, Vardi Y. Maximum likelihood reconstruction for emission tomography. IEEE transactions on medical imaging. 1982;1(2):113–122. pmid:18238264
- 9. Nuyts J, De Man B, Dupont P, Defrise M, Suetens P, Mortelmans L. Iterative reconstruction for helical CT: a simulation study. Physics in Medicine & Biology. 1998;43(4):729.
- 10.
Buades A, Coll B, Morel JM. A non-local algorithm for image denoising. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05). vol. 2. IEEE; 2005. p. 60–65.
- 11.
Dabov K, Foi A, Katkovnik V, Egiazarian K. Image denoising with block-matching and 3D filtering. In: Image Processing: Algorithms and Systems, Neural Networks, and Machine Learning. vol. 6064. International Society for Optics and Photonics; 2006. p. 606414.
- 12.
Yu S, Muhammed HH. Noise type evaluation in positron emission tomography images. In: 2016 1st International Conference on Biomedical Engineering (IBIOMED). IEEE; 2016. p. 1–6.
- 13. Dutta J, Leahy RM, Li Q. Non-local means denoising of dynamic PET images. PloS one. 2013;8(12):e81390. pmid:24339921
- 14. Chen H, Zhang Y, Zhang W, Liao P, Li K, Zhou J, et al. Low-dose CT via convolutional neural network. Biomedical optics express. 2017;8(2):679–694. pmid:28270976
- 15. Chen H, Zhang Y, Kalra MK, Lin F, Chen Y, Liao P, et al. Low-dose CT with a residual encoder-decoder convolutional neural network. IEEE transactions on medical imaging. 2017;36(12):2524–2535. pmid:28622671
- 16.
Wu D, Kim K, Fakhri GE, Li Q. A cascaded convolutional neural network for x-ray low-dose CT image denoising. arXiv preprint arXiv:170504267. 2017;.
- 17. Wolterink JM, Leiner T, Viergever MA, Išgum I. Generative adversarial networks for noise reduction in low-dose CT. IEEE transactions on medical imaging. 2017;36(12):2536–2545. pmid:28574346
- 18. Yang Q, Yan P, Zhang Y, Yu H, Shi Y, Mou X, et al. Low-dose CT image denoising using a generative adversarial network with Wasserstein distance and perceptual loss. IEEE transactions on medical imaging. 2018;37(6):1348–1357. pmid:29870364
- 19. Xiang L, Qiao Y, Nie D, An L, Lin W, Wang Q, et al. Deep auto-context convolutional neural networks for standard-dose PET image estimation from low-dose PET/MRI. Neurocomputing. 2017;267:406–416. pmid:29217875
- 20.
Xu J, Gong E, Pauly J, Zaharchuk G. 200x low-dose pet reconstruction using deep learning. arXiv preprint arXiv:171204119. 2017;.
- 21. Wang Y, Yu B, Wang L, Zu C, Lalush DS, Lin W, et al. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage. 2018;174:550–562. pmid:29571715
- 22. Kaplan S, Zhu YM. Full-Dose PET Image Estimation from Low-Dose PET Image Using Deep Learning: a Pilot Study. Journal of digital imaging. 2018; p. 1–6.
- 23. Kawahara J, Brown CJ, Miller SP, Booth BG, Chau V, Grunau RE, et al. BrainNetCNN: convolutional neural networks for brain networks; towards predicting neurodevelopment. NeuroImage. 2017;146:1038–1049. pmid:27693612
- 24.
He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision; 2015. p. 1026–1034.
- 25.
Taigman Y, Yang M, Ranzato M, Wolf L. Deepface: Closing the gap to human-level performance in face verification. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2014. p. 1701–1708.
- 26.
Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Springer; 2015. p. 234–241.
- 27. Chen KT, Gong E, de Carvalho Macruz FB, Xu J, Boumis A, Khalighi M, et al. Ultra–Low-Dose 18F-Florbetaben Amyloid PET Imaging Using Deep Learning with Multi-Contrast MRI Inputs. Radiology. 2018;290(3):649–656. pmid:30526350
- 28. Ouyang J, Chen KT, Gong E, Pauly J, Zaharchuk G. Ultra-low-dose PET Reconstruction using Generative Adversarial Network with Feature Matching and Task-Specific Perceptual Loss. Medical physics. 2019;.
- 29. Gong K, Guan J, Kim K, Zhang X, Yang J, Seo Y, et al. Iterative PET image reconstruction using convolutional neural network representation. IEEE transactions on medical imaging. 2018;38(3):675–685.
- 30. Haggstrom I, Schmidtlein CR, Campanella G, Fuchs TJ. DeepPET: A deep encoder–decoder network for directly solving the PET image reconstruction inverse problem. Medical image analysis. 2019;54:253–262. pmid:30954852
- 31.
Feuardent J, Soret M, De Dreuille O, Foehrenbach H, Buvat I. Reliability of SUV estimates in FDG PET as a function of acquisition and processing protocols. In: 2003 IEEE Nuclear Science Symposium. Conference Record (IEEE Cat. No. 03CH37515). vol. 4. IEEE; 2003. p. 2877–2881.
- 32. Kinahan P, Fletcher J. PET/CT standardized uptake values (SUVs) in clinical practice and assessing response to therapy. Semin Ultrasound CT MR. 2010;31(6):496–505. pmid:21147377
- 33. Lindholm H, Staaf J, Jacobsson H, Brolin F, Hatherly R, Sânchez-Crespo A. Repeatability of the maximum standard uptake value (SUVmax) in FDG PET. Molecular imaging and radionuclide therapy. 2014;23(1):16. pmid:24653930
- 34.
Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2017. p. 4700–4708.
- 35.
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.
- 36.
Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision; 2017. p. 2223–2232.
- 37. Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC. Improved training of wasserstein gans. In: Advances in Neural Information Processing Systems; 2017. p. 5767–5777.
- 38. Kang E, Koo HJ, Yang DH, Seo JB, Ye JC. Cycle-consistent adversarial denoising network for multiphase coronary CT angiography. Medical physics. 2019;46(2):550–562. pmid:30449055
- 39. Kang E, Chang W, Yoo J, Ye JC. Deep convolutional framelet denosing for low-dose CT via wavelet residual network. IEEE transactions on medical imaging. 2018;37(6):1358–1369. pmid:29870365
- 40. Schaefferkoetter J, Nai YH, Reilhac A, Townsend DW, Eriksson L, Conti M. Low dose positron emission tomography emulation from decimated high statistics: A clinical validation study. Medical physics. 2019;. pmid:30929270
- 41. Jan S, Santin G, Strul D, Staelens S, Assie K, Autret D, et al. GATE: a simulation toolkit for PET and SPECT. Physics in Medicine & Biology. 2004;49(19):4543.
- 42. Jan S, Benoit D, Becheva E, Carlier T, Cassol F, Descourt P, et al. GATE V6: a major enhancement of the GATE simulation platform enabling modelling of CT and radiotherapy. Physics in Medicine & Biology. 2011;56(4):881.
- 43. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP, et al. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing. 2004;13(4):600–612. pmid:15376593
- 44.
Zhang R, Isola P, Efros AA, Shechtman E, Wang O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT: IEEE; 2018. p. 586–595. Available from: https://ieeexplore.ieee.org/document/8578166/.
- 45.
Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014;.
- 46.
Chollet F, et al.. Keras; 2015. https://keras.io.
- 47.
Zhang R, Isola P, Efros AA, Shechtman E, Wang O. The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2018. p. 586–595.
- 48. Lei Y, Dong X, Wang T, Higgins K, Liu T, Curran WJ, et al. Whole-body PET estimation from low count statistics using cycle-consistent generative adversarial networks. Physics in Medicine & Biology. 2019;64(21):215017.
- 49. Soret M, Bacharach SL, Buvat I. Partial-Volume Effect in PET Tumor Imaging. The Journal of Nuclear Medicine. 2007;48(6):932–945. pmid:17504879