Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Semi-supervised synthesis of 7T MRI from 3T using 3D FR-U-Net with anatomical segmentation consistency assessment

  • Richard Acs ,

    Roles Conceptualization, Data curation, Investigation, Methodology, Software, Writing – original draft, Writing – review & editing

    racs2019@fau.edu

    Affiliation College of Engineering and Computer Science, Florida Atlantic University, Boca Raton, Florida, United States of America

  • Hanqi Zhuang

    Roles Conceptualization, Methodology, Supervision, Writing – review & editing

    Affiliation College of Engineering and Computer Science, Florida Atlantic University, Boca Raton, Florida, United States of America

Abstract

Ultra-high-field 7T MRI provides substantial benefits for neuroimaging, including improved resolution and contrast, but remains limited by high costs and restricted accessibility. In this study, we propose a semi-supervised 3D multi-scale fusion residual U-Net (semi supervised 3D FR-U-Net) to synthesize 7T MRI volumes from 3T input using a patch-based architecture optimized for low-data settings. In addition to evaluating conventional synthesis metrics such as PSNR, SSIM, and NMSE, we introduce a novel segmentation-based assessment using the VolBrain pipeline to quantify anatomical fidelity. Our model outperforms prior methods—even without preprocessing steps like skull stripping—and achieves high fidelity in global brain morphology and basal ganglia structures. However, significant asymmetry errors in hippocampal segmentation highlight limitations in preserving fine, clinically critical anatomy. To address the disconnect between technical performance and clinical applicability, we emphasize the use of interpretable, segmentation-derived metrics to bridge the gap between research advances in synthetic MRI and real-world diagnostic relevance. These findings underscore the importance of region-specific evaluation and demonstrate how structural metrics can guide the real-world applicability of synthetic MRI, particularly when expert radiological review is not feasible.

Introduction

Magnetic resonance imaging (MRI) is one of the most widely used medical imaging modalities, offering high-resolution, non-invasive visualization of soft tissues. Unlike computed tomography (CT) or X-ray imaging, which rely on ionizing radiation, MRI leverages strong magnetic fields and radiofrequency pulses to generate detailed anatomical and functional images [1]. This versatility has made MRI an essential tool in various clinical and research applications, including neurology, oncology, cardiology, and musculoskeletal imaging [24]. The quality and diagnostic power of an MRI scan are influenced by several factors, one of the most critical being the strength of the magnetic field, measured in tesla (T) [5].

Higher-field MRI scanners, such as ultra-high-field 7T systems, provide substantial advantages over standard 3T and 1.5T MRIs, offering improved signal-to-noise ratio (SNR), enhanced tissue contrast, and greater spatial resolution [5]. These benefits are particularly significant in neuroimaging, where 7T MRI enables the visualization of fine-scale structures such as cortical layers, hippocampal subfields, and small subcortical nuclei, supporting more precise structural analysis and improved lesion detection [6]. As a result, 7T has become an important tool for selected clinical and research applications—especially in the study of neurodegenerative diseases, epilepsy, and multiple sclerosis [7,8]. However, 7T imaging also introduces significant technical challenges, including B0 and B1 field inhomogeneities, increased susceptibility to artifacts, and higher specific absorption rate (SAR), which can limit performance in certain sequences [6]. For instance, while 7T enhances contrast in susceptibility-weighted and T2*-weighted imaging, it may produce suboptimal results in sequences like T1-weighted MPRAGE without careful protocol optimization. Therefore, 7T does not consistently outperform 3T across all imaging contexts, and its advantages depend heavily on the imaging task, protocol, and anatomical target.

In addition to these technical considerations, the widespread adoption of 7T remains constrained by high cost, limited availability, and logistical barriers [5]. This makes it a valuable but resource-intensive modality, best reserved for cases where its benefits clearly outweigh these constraints. In many routine or large-scale imaging scenarios—particularly when the expected diagnostic gain is marginal—acquiring a full 7T scan may not be justified. In such contexts, algorithmic synthesis of 7T-equivalent images from more accessible 3T scans presents a promising alternative.

Given these challenges, there is strong motivation to develop deep learning-based approaches for synthesizing high-field MRI from lower-field acquisitions, particularly for translating 3T MRI into 7T-equivalent images. By generating high-quality 7T-like scans from more widely available 3T data, such approaches have the potential to enhance clinical imaging, improve diagnostic precision, and expand the utility of ultra-high-field imaging without requiring direct access to 7T scanners [9]. Additionally, the ability to synthesize 7T images from large-scale 3T datasets can facilitate augmentation of existing 7T data, enabling the development of more robust machine learning models for disease classification, segmentation, and other downstream tasks [9].

Recent advances in deep learning, particularly generative models such as generative adversarial networks (GANs), diffusion models, and transformer models, have demonstrated promise in medical image synthesis [9]. Prior studies on high-field MRI synthesis have explored various methodologies, including slice-by-slice synthesis, full-volume synthesis, and hybrid patch-based approaches [912]. While slice-wise synthesis methods are computationally efficient and require less memory, they often introduce inconsistencies between adjacent slices, leading to volumetric discontinuities that degrade anatomical fidelity [9]. Full-volume synthesis methods, on the other hand, provide more structurally coherent reconstructions by accounting for inter-slice dependencies, but they typically require larger datasets and substantial computational resources [9].

A major limitation in the field of 7T MRI synthesis is the scarcity of publicly available datasets. Unlike 3T and 1.5T MRI, for which numerous large-scale datasets exist, high-field (≥7T) MRI datasets are limited, with most being private or containing only a small number of subjects [13]. This lack of data presents a significant challenge for training deep learning models, as large datasets are often necessary to achieve generalizable and high-fidelity synthesis. Currently, there are an estimated 100 7T scanners in use, making the development of new datasets mostly limited to small single-institutional cohorts [14]. In this study, we aim to explore volumetric synthesis methods using a recently curated dataset of 10 paired 3T and 7T T1-weighted MRI volumes [15]. More specifically, we present a semi-supervised 3D multi-scale fusion residual U-Net (semi-supervised 3D FR-U-Net) to synthesize full 7T MRI volumes from input 3T volumes. While U-Net architectures have been widely used in MRI synthesis [9], recent advances in diffusion models (DDPMs), vision transformers, and GANs have set new benchmarks in image synthesis quality. However, these models often require large-scale datasets and extensive computational resources, making them impractical for low-resource settings. Our approach retains the computational and data efficiency of U-Nets while integrating residual block structures and multiscale fusion—techniques inspired by high performance segmentation models and vision transformers—to improve representational capacity and cross-scale consistency in MRI synthesis. Additionally, we leverage unpaired 7T data through semi-supervised learning to enhance anatomical fidelity of synthetic high-field MRI structure. Previous approaches to MRI synthesis have used semi-supervised learning [16,17]. However, our study is the first to merge semi-supervised learning into a 3D-patch based U-net approach. Thus, we demonstrate a scalable and efficient alternative for MRI synthesis when paired data is limited, a common challenge in real-world clinical applications of ultra-high-field (7T) MRIs.

This work presents several key contributions to the field of ultra-high-field MRI synthesis under small-data constraints. We utilize a recently curated public dataset consisting of only 10 paired 3T and 7T T1-weighted MRI volumes, making this one of the smallest-scale studies to date to perform direct full-volume synthesis rather than conventional slice-wise approaches. While prior studies on this dataset have focused exclusively on 2D synthesis methods [17,18], and Cui et al. applied 3D patches only after skull stripping and harmonization for microbleed synthesis [14], our study is the first to conduct full-volumetric synthesis from minimally preprocessed 3T data. To maximize data efficiency and generalization, we adopt a 3D patch-based U-Net architecture and further enhance anatomical fidelity through a semi-supervised learning framework that leverages unpaired 7T scans for domain adaptation.

In addition to this novel synthesis pipeline, we introduce an anatomical segmentation evaluation framework based on VolBrain, an automated brain segmentation tool built on a large ensemble of CNNs [19]. Beyond standard Dice scores, we propose segmentation consistency derived metrics that capture region-specific anatomical accuracy and asymmetry. This structure-based analysis provides valuable insights into clinical utility without requiring manual annotation when radiologist review is unavailable. Our contributions are summarized as follows:

  1. Demonstrating the feasibility of volumetric ultra-high-field MRI synthesis with extremely small datasets (10 paired scans).
  2. Presenting a novel semi-supervised 3D multi-scale fusion residual U-Net (semi-supervised 3D FR-U-Net).
  3. Enhancing anatomical fidelity through semi-supervised learning, leveraging unpaired data for improved generalization.
  4. Utilizing segmentation-based volumetric fidelity metrics to provide interpretable, anatomically-relevant evaluations beyond conventional pixel-level metrics.

Related approaches

The task of synthesizing 7T MRI from 3T images is best characterized as a cross-domain image translation problem rather than a traditional super-resolution task. While 7T MRI often provides higher spatial resolution, the more critical challenge lies in modeling the differences in contrast, signal properties, and anatomical representation between the two field strengths. Early approaches relied on non-deep learning techniques. For instance, Bahrami et al. [20] employed a random forest-based regression model to map 3T images to 7T, while another study utilized a canonical correlation analysis (CCA)-based approach for synthesis [21]. Although these methods established initial benchmarks, they struggled to preserve structural similarity and anatomical fidelity, limiting their clinical applicability.

With the rise of deep learning, CNN-based and generative models have become the dominant methods for high-field MRI synthesis. Bahrami et al. [22] introduced a deep CNN-based approach, which outperformed previous traditional techniques but suffered from limited generalization due to small dataset sizes. More advanced generative models have since been applied to high-field MRI synthesis. Wu et al. demonstrated that diffusion probabilistic models (DDPMs) can achieve high structural similarity and anatomical fidelity while reducing image distortion [23]. Similarly, conditional GANs have been used to synthesize 7T images from 3T counterparts, achieving improvements in image quality, artifact suppression, sharpness, and contrast, as assessed through manual radiological Likert scale evaluations [24].

However, these deep learning approaches typically require large datasets, which pose a significant challenge given the scarcity of 7T scanners. For example, the DDPM method used 928 MRI subjects to generate a paired low-high resolution dataset [23], while a cycle GAN-based method relied on a privately curated dataset of 122 paired 3T and 7T MRI volumes [24]. Through recent literature, most paired 3T-7T studies have been conducted using as few as 15 patient scans, with larger studies restricted to private datasets, limiting model generalization [13,25].

To address these data limitations, researchers have developed techniques for improving deep learning model performance with small datasets. One approach involved integrating wavelet-based affine transformation layers, achieving high structural similarity and PSNR performance on just 15 paired MRI volumes [25]. Another strategy, Unlimited Data Augmentation (UDA), leveraged deformable registration to artificially expand a 15-pair dataset, integrating a cycle GAN for 7T synthesis [13]. Additionally, various deep learning architectures have been explored for small datasets, including V-Net-based models [14], cycle GAN approaches [18], and 3D anisotropic U-Net models for 1.5 to 3T synthesis [26,27]. Another study employed a wavelet-based frequency attention network with semi-supervised dual-domain consistency learning to synthesize 7T slices from 3T inputs [17].

Most existing studies focus on 2D slice-based synthesis, which has lower data requirements but sacrifices anatomical consistency across slices. In contrast, 3D synthesis improves anatomical coherence but demands significantly larger datasets. Of the literature reviewed, two studies utilized volumetric inputs and outputs, a V-Net-based 3D patch approach [14], and 3D patch-based U-Net approach [28]. Such studies are typically accompanied by dice scores against a segmentation ground truth to evaluate their volumetric fidelity in specific structural areas, such as hippocampal subfields [28]. In a recent 2D-based approach, Eidex et al. [29] introduced the 7T-Restormer, an efficient transformer-based model capable of synthesizing quantitative 7T relaxometry maps from combined 1.5T and 3T inputs, showing improved generalization across field strengths. However, their transformer based approach, even when focused on efficiency, required over 100 training MRI volumes,typical of state of the art architecture. Thus, U-Net variations and GAN-based models with data augmentation strategies remain dominant in small-sample MRI synthesis studies, both on 2D and 3D data.

In summary, prior work is often limited by dependence on large or private datasets, the use of predominantly 2D architectures that introduce inter-slice discontinuities, and the absence of region-specific anatomical evaluation beyond global similarity metrics. Many also require extensive preprocessing (e.g., skull stripping), which can limit real-world applicability. This study addresses these issues by enabling volumetric synthesis in a low-data setting, leveraging unpaired 7T data to improve generalization, minimizing preprocessing, and incorporating segmentation-based metrics for anatomically relevant assessment.

Materials and methods

This study investigates the synthesis of 7T MRI volumes from 3T inputs through several GAN and U-net based approaches. In these methods, either individual 2D slices or 3D patches are synthesized and subsequently assembled to reconstruct the final volumetric image. The quality of the synthesized volumes is assessed through both quantitative metrics (PSNR, SSIM, NMSE) and segmentation-based volumetric analysis. All experiments were conducted on an NVIDIA RTX A6000 GPU (48 GB memory), 8-core CPU, and 45 GB system RAM, running Python 3.13.5 with TensorFlow 2.16.1, CUDA 12.2, and cuDNN 8.9 on Ubuntu 22.04.

All data were preprocessed following the protocol in Sect ‘Data pre-processing’, ensuring consistency across training and evaluation. We then trained and compared three models, a 3D multi-scale fusion residual U-Net (3D FR-U-Net), a semi-supervised 3D multi-scale fusion residual U-Net (semi-supervised 3D FR-U-Net), and a 2D Slice-by-Slice Pix2Pix Baseline. An ablation study was also conducted to demonstrate the effectiveness of the components of our semi-supervised 3D FR-U-Net.

Each model was evaluated using 10-fold cross-validation, and performance was measured using PSNR, SSIM, and NMSE, as detailed in ’Quantitative Evaluation Metrics’. Additionally, we conducted segmentation analysis on the best-performing model to assess structural consistency, as described in ’Volumetric Evaluation Metrics’. The following sections provide a detailed breakdown of the dataset, model architectures, and evaluation methodology.

Datasets

In this study, we utilize the 7T and 3T Paired Brain MRI Dataset published by Chen et al. [15], which consists of paired MRI scans acquired at both 3T and 7T field strengths. The dataset comprises 10 subjects, each with a 3T and a 7T T1-weighted and T2-weighted MRI volume for a total of 4 volumes per subject. The 3T scans were acquired using a Siemens MAGNETOM Prisma 3T scanner, while the 7T scans were obtained using a Siemens MAGNETOM Terra 7T scanner. Both scans were acquired with high isotropic resolution and underwent preprocessing to ensure alignment between the 3T and 7T images. The dataset includes co-registered images, allowing for voxel-wise comparisons between the real and synthesized 7T images. The 7T T1-weighted MP2RAGE sequence acquires two images at different inversion times, each using a separate flip angle (FA1 and FA2), which are then combined to produce a single high-contrast, bias-corrected T1-weighted image. These flip angles are not alternated during acquisition but correspond to distinct readouts that enhance T1 contrast and reduce B1+ inhomogeneity effects, which is particularly important at 7T field strength [6]. Table 1 shows key acquisition characteristics of the different scan types.

thumbnail
Table 1. Acquisition characteristics of 7T and 3T MRI images. All scans were acquired using the MPRAGE protocol.

https://doi.org/10.1371/journal.pone.0333499.t001

The dataset provides both volumes with the original characteristics and aligned versions in which the 3T T1w images were linearly registered to the 7T T1w images. Linear registration was performed via rigid-body alignment with translations and rotations, without scaling or nonlinear deformation, using FSL FLIRT to ensure anatomical correspondence between the 3T and 7T scans [15]. Due to model architecture considerations, and to make direct comparison of synthesized samples easier, we utilized the aligned versions in this study. Thus, all volumes contain 256 sagittal, 304 coronal, and 308 transverse slices. Fig 1 shows the different scan types provided by the paired dataset. Aside from linear registration and facial information removal from all MRI scans, the dataset utilized for this study had no other preprocessing applied.

thumbnail
Fig 1. Different MRI scan contrasts and strengths provided by the paired dataset [15].

https://doi.org/10.1371/journal.pone.0333499.g001

For our semi-supervised approach, we employ an additional dataset of 20 7T volumes presented by Li et al. [28]. The dataset contains T1-weighted whole-brain scans from 20 healthy university volunteers (10 males, 10 females) aged 18–25 years, collected at the Beijing MRI Center for Brain Research. Scans were acquired on the same day using an investigational Nova 7T scanner. The 7T protocol included a high-resolution 3D T2-weighted SPACE sequence (0.4 × 0.4 × 1 mm3) and an isotropic 0.7 mm3 T1-weighted MPRAGE sequence, both scanned along the hippocampal long axis and the AC-PC line. We only utilize the 7T T1-weighted data for our semi-supervised approach since we focus only on high-field synthesis of T1-weighted samples. Each 7T volume contains an acquisition matrix of size 320 × 320 × 256, resulting in a total of 11340 additional 64 x 64 x 64 patches to leverage for our semi-supervised learning approach.

All participants in the Chen et al. [15] and Li et al. [28] datasets provided informed written consent for data collection and sharing. The data are fully anonymized, and no identifiable personal information was used in this study. The original studies that collected these data were approved by their respective institutional review boards (IRBs) at their hosting institutions. As we only used anonymized, previously collected MRI data without any personal identifiers or interaction with subjects, additional IRB approval was not required at Florida Atlantic University. The datasets were accessed on February 15, 2025, and at no point did the authors have access to information that could identify individual participants. All data use complies with institutional and journal ethical standards.

Data pre-processing

For this study, each MRI volume was normalized to a range of [0,1] before training/testing, ensuring standardization while maintaining subject level intensity distributions. We utilize a 10-fold cross validation technique in which one full MRI volume is used for testing, while the remaining 9 are used for training, following the leave one out cross validation approach utilized by previous studies [22,25,30]. Next, for the patch-based method, each of the volumes were split into 64 x 64 x 64 patches with an overlapping stride of 32. This resulted in 567 patches per volume, for a total of 5103 paired patches in the training set. For the slice-by-slice synthesis methods, we extracted all paired slices for each direction, resulting in 8,800 paired slices for training and testing. Once again, we utilized 10-fold cross validation and set aside a full MRI volume for testing during each fold. Fig 2 illustrates the overall preprocessing workflow that we used for this study.

thumbnail
Fig 2. Dataset preprocessing workflow for paired T1W 7T and 3T MRI data [15].

https://doi.org/10.1371/journal.pone.0333499.g002

3D U-Net based models

To address the challenge of volumetric medical image translation, we propose a 3D U-Net inspired architecture designed for learning fine-grained structural features while maintaining global spatial consistency. The architecture follows an encoder-decoder paradigm with symmetric skip connections, allowing multi-scale feature integration and efficient high-resolution 3D image reconstruction. Fig 3 illustrates the overall architecture of our 3D Multi-Scale Fusion Residual U-Net. Our architecture is based on the U-Net presented by Li et al. [28], with key modifications such as attention through residual blocks and multi-scale fusion.

thumbnail
Fig 3. 3D multi-scale fusion residual U-Net (3D FR-U-Net) architecture.

https://doi.org/10.1371/journal.pone.0333499.g003

The encoder consists of four hierarchical levels, each composed of 3D convolutional layers with parametric rectified linear unit (PReLU) activations, followed by layer normalization for stable gradient propagation. A max-pooling operation with a stride of (2 × 2 × 2) is applied at each level to downsample spatial resolution while increasing feature dimensionality. The encoder stages are as follows:

  1. 3D convolutional layers with 32 feature channels; spatial resolution 64 × 64 × 64
  2. 3D convolutional layers with 64 feature channels; spatial resolution 32 × 32 × 32
  3. 3D convolutional layers with 128 feature channels; spatial resolution 16 × 16 × 16
  4. (Bottleneck) 3D convolutional layers with 256 feature channels; spatial resolution 8 × 8 × 8

The decoder reconstructs the volumetric MRI using 3D transposed convolutions (deconvolutions) with residual blocks and multi-scale fusion. The transposed convolutions has a stride of (2 × 2 × 2) at each level to up sample spatial resolution while reducing feature dimensionality. The decoder stages are as follows:

  1. Transposed convolution (128 channels) and concatenation with encoder level 3
  2. Transposed convolution (64 channels) and concatenation with encoder level 2
  3. Transposed convolution (32 channels) and concatenation with encoder level 1
  4. Final Reconstruction: convolution followed by a sigmoid activation function

Residual blocks are introduced at each decoding stage to refine feature maps [31], ensuring better anatomical accuracy. Each residual block consists of two (3 × 3 × 3) convolutional layers with PReLU activations, and a skip connection that adds the input directly to the output. To capture fine details across different receptive fields, we integrate a multi-scale feature extraction strategy [31]. At each decoder stage, three parallel 3D convolutional branches process the input feature maps using (1 × 1 × 1), (3 × 3 × 3), and (5 × 5 × 5) kernels, followed by channel-wise concatenation.

To optimize training, we employ a hybrid loss function that combines Mean Squared Error (MSE) for intensity-based reconstruction and Structural Similarity Index Measure (SSIM) to enforce perceptual similarity. The hybrid loss function is defined by Eq 1, where λ is a weighting factor that controls the influence of SSIM loss. We set λ = 0.7 to emphasize perceptual quality while maintaining voxel intensity consistency.

(1)

Eq 4 defines the Structural Similarity Index Measure (SSIM) between an image and a target image, where and represent the mean voxel intensities, and denote the variances, is the covariance, and C1 and C2 are stabilizing constants. SSIM loss, shown in Eq 2, calculates the mean SSIM score across all N voxel-wise patches or slices. In contrast, mean squared error (MSE) loss, shown in Eq 3, computes the average squared difference between the predicted voxel intensities and the ground truth voxel intensities yi over N samples.

(2)(3)(4)

The model is trained using the Adam optimizer with a learning rate of . A batch size of 8 is used due to memory constraints, and training is conducted for 100 epochs. To prevent overfitting, early stopping is applied based on the validation loss, with a patience of 10 epochs.

Model ablation

To evaluate the effect of each architectural modification on synthesis performance, we conducted an ablation study based on a standard 3D U-Net. Starting with the baseline model, we incrementally introduced the following components: (1) reduced network depth by one layer to mitigate overfitting, (2) a hybrid loss function combining SSIM and MSE, (3) residual blocks and multi-scale fusion (MSF) modules in the decoder, and (4) semi-supervised training. Each addition was assessed to isolate its individual contribution to the overall performance.

Semi-supervised training

In this study, we implemented a semi-supervised learning framework to enhance the translation of synthetic 3T MRI patches to high field 7T MRI. Our approach leverages both paired (3T  7T) data for supervised learning and unpaired 7T MRI patches to enforce consistency constraints on the model’s predictions. Specifically, we utilized the 3D Multi-Scale Fusion Residual U-Net as our backbone model as presented in the previous sections, trained using a hybrid loss function that combines Mean Squared Error (MSE) and Structural Similarity Index (SSIM) for paired data. To incorporate the unpaired 7T data, we introduced a consistency loss, which encourages the network to produce structurally coherent outputs by minimizing L1 and SSIM differences between predicted and real 7T images.

The consistency loss for unpaired 7T MRI patches is defined by Eq 5, where x7T represents the real unpaired 7T MRI patch, f(x7T) is the corresponding output generated by the model, and SSIM measures the structural similarity between the input and output. The hyperparameter α controls the contribution of SSIM loss to ensure structural coherence.

(5)

The total loss function used for training is defined in Eq 6, where denotes the hybrid loss computed for paired () samples, and λ is a weighting factor that controls the influence of the consistency loss during training.

(6)

The dataset was structured to maintain a 1:1 ratio of unpaired to paired samples, ensuring effective regularization without over-constraining the model. Fig 4 shows the workflow of the semi-supervised training process. During each batch, the loss was individually calculated for the paired and unpaired samples after a forward pass through the model, resulting in a combined loss which was used to update the model. This was repeated until early stopping convergence with patience of 10 epochs was achieved or 100 epochs total. Training was conducted using a custom TensorFlow implementation with a subclassed Keras model, where the semi-supervised optimization was performed via a modified ‘train step()‘ function to jointly compute gradients from both paired and unpaired loss components. Model evaluation was performed exclusively on the supervised validation dataset, ensuring that performance comparisons remained consistent with fully supervised baselines.

Slice based methods

As a benchmark synthesis method, we use the pix2pix conditional GAN architecture extended with ResNet blocks to introduce elements of attention. Pix2pix has been widely used for medical image synthesis, including MRI synthesis, and is utilized as a benchmark GAN by many studies [9,11,12]. We follow the implementation details presented by [32], only modifying the generator architecture to add 3 ResNet blocks between the down sampling and up sampling layers. We did this due to introduce a basic form of attention, which we also utilized in the 3D volumetric models and allowed us to achieve an improved baseline performance.

We selected pix2pix as a baseline due to its widespread adoption in medical image synthesis literature especially in data-constrained scenarios. While more advanced models such as DDPMs and transformer-based architectures exist, they typically require large datasets and extensive tuning. These models were excluded due to incompatibility with our small-data setting and the unavailability of pretrained implementations suitable for volumetric MRI synthesis.

Quantitative evaluation metrics

For our evaluation metrics, we used three main qualitative metrics, structural similarity index measure (SSIM), peak signal-to-noise ratio (PSNR), and normalized mean squared error (NMSE), which are commonly used in conditional image synthesis tasks. Eq 4, Eq 7, and Eq 8 represent PSNR, NMSE, and SSIM respectively.

(7)(8)

In the PSNR Eq 7, y(x) represents the ground truth image, while G(x) is the reconstructed image. The numerator corresponds to the maximum possible pixel value, while the denominator contains the Mean Squared Error (MSE), which quantifies the difference between the two images. The NMSE Eq 8 normalizes the squared error between y(x) and G(x) by the squared norm of the ground truth image, providing a scale-invariant measure of error. The SSIM Eq 4 evaluates structural similarity by considering luminance, contrast, and structural details between the images. It incorporates mean intensity, variance, and covariance terms to measure perceptual quality while stabilizing computations with predefined constants.

To gain a more comprehensive understanding of the synthesis quality of the MRI volume, we report the above metrics separately for each slice orientation of the synthetic MRI samples. Since MRI volumes can be viewed from multiple axes, each providing a distinct perspective on anatomical structures, evaluating synthesis quality across all orientations allows us to identify any discrepancies. This analysis helps determine if certain orientations exhibit lower synthesis fidelity.

Qualitative evaluation metrics

Aside from standard qualitative metrics, we incorporate segmentation-based volumetric evaluation to assess the structural fidelity of the synthesized 7T MRI volumes. Specifically, we utilize VolBrain [19], a high-resolution neuroimaging segmentation pipeline known for its anatomical precision in fine-grained whole brain structure delineation. VolBrain is based on AssemblyNet, a large ensemble of 3D convolutional neural networks (CNNs) trained to segment 132 brain structures at high resolution using a multi-scale, two-assembly U-Net architecture [19]. Table 2 describes the regional volumetric areas of interest analyzed using VolBrain, which serve as key reference points for our assessment. We assess anatomical consistency based on medically established associations between structural changes in each region and specific neurological or psychiatric conditions [3336].

thumbnail
Table 2. Volumetric segmentation masks and clinical relevance.

https://doi.org/10.1371/journal.pone.0333499.t002

By comparing segmentation outputs between real and synthetic 7T MRIs, we can quantify structural differences and assess the model’s ability to preserve detailed anatomical features. VolBrain performs voxel-wise segmentation across an anatomical hierarchy, providing a comprehensive breakdown of each MRI volume into labeled subregions. We utilize these anatomical segmentations to quantify the discrepancy between ground truth and synthetic outputs, capturing volumetric deviations across specific regions. We compute segmentation-derived metrics, including the Dice Similarity Coefficient (DSC) and Hausdorff Distance (HD), to measure overlap accuracy and boundary precision between real and synthetic 7T segmentations. DSC and HD are widely used to assess segmentation performance in medical imaging [28]. DSC quantifies how similar two sets are by comparing the number of overlapping voxels to the total number of voxels in both segmentations, as described by Eq 9.

(9)

The Hausdorff Distance (HD) is a boundary-based metric used to quantify the spatial discrepancy between two segmentation outputs. It measures the maximum distance from a point in one segmentation to the closest point in the other. Formally, given two sets of points A and B, the directed Hausdorff distance from A to B is defined by Eq 10, and the symmetric Hausdorff Distance is then calculated as shown in Eq 11.

(10)(11)

This metric captures the worst-case scenario of mismatch between boundaries, making it sensitive to outliers. Therefore, we will use the 95th percentile Hausdorff Distance (HD95) to provide a more robust assessment by excluding extreme outlier deviations.

Results

This section presents the results of the experiments outlined in the Methods section, focusing on the evaluation of synthesized 7T MRI volumes. The results are analyzed using quantitative metrics such as PSNR, NMSE, and SSIM, as well as segmentation-derived evaluations including Dice Similarity Coefficient (DSC) and Hausdorff Distance. Additionally, qualitative visual comparisons provide further insights into the anatomical fidelity of the generated images.

Synthesis quality

Table 3 presents the leave-one-out performance by slice orientation from the ablation study on the components of our final semi-supervised 3D FR-U-Net architecture. Subject 8 was used for validation, while the remaining 9 subjects were used for training. As shown, each architectural modification progressively improves structural similarity and other performance metrics, validating the contributions of each component.

thumbnail
Table 3. Ablation study of the semi-supervised 3D FR-U-Net architecture.

https://doi.org/10.1371/journal.pone.0333499.t003

Table 4 shows the 10-fold cross-validation results by slice orientation for the three generative models evaluated in this study. The best-performing metrics for each orientation are highlighted in bold. Among the models tested, the semi supervised 3D FR-U-Net achieved the highest overall in SSIM metrics.

thumbnail
Table 4. Mean ± standard deviation of PSNR, SSIM, and NMSE by slice orientation for different generative models.

https://doi.org/10.1371/journal.pone.0333499.t004

To assess the effectiveness of our semi-supervised 3D FR-U-Net, we conducted paired t-tests comparing its performance to two baselines: Pix2Pix and a fully supervised 3D FR-U-Net. The semi-supervised 3D FR-U-Net significantly outperformed Pix2Pix in all slice orientations for SSIM and PSNR (all p < 0.0001), and also achieved significantly lower NMSE in the transverse (p = 0.0228) and sagittal (p = 0.0135) planes, with coronal NMSE showing a non-significant trend toward improvement (p = 0.066). In contrast, no statistically significant differences were found when comparing the semi-supervised FR-U-Net with the supervised FR-U-Net, across any metric or orientation (all p > 0.28). The corresponding t-values for each comparison are summarized in Table 5.

thumbnail
Table 5. Paired t-test statistics (t-values) comparing the semi-supervised FR-U-Net to baseline models across image quality metrics and slice orientations.

https://doi.org/10.1371/journal.pone.0333499.t005

Anatomical segmentation

Fig 5 shows the average Dice Scores by segmentation map types between the ground truth T1W 7T and synthetic T1W 7T MRI volumes. VolBrain produces 5 different segmentation masks, including a binary mask with 1 class, a lobes mask with 12 classes, a macrostructure mask with 6 classes, a structures class with 132 classes, and a tissues mask with 7 classes. Fig 6 shows the average Hausdorff-95 distance between the ground truth and synthetic 7T MRI volumes for the same 5 segmentation mask types.

thumbnail
Fig 5. Segmentation Dice scores across anatomical categories for synthetic T1W 7T volumes.

Segmentation dice scores are reported as average across all 10-fold test volumes from the paired dataset by Chen et al. [15]. Dice similarity coefficients are shown for five segmentation classes using VolBrain: (1) binary brain mask (brain tissue vs. non-tissue), (2) lobar segmentation based on functional divisions, (3) macrostructures representing coarse anatomical groupings, (4) finer anatomical structures, and (5) primary tissue classes.

https://doi.org/10.1371/journal.pone.0333499.g005

thumbnail
Fig 6. Hausdorff distances across anatomical segmentation categories for synthetic 7T volumes.

Lower values indicate better anatomical boundary alignment between synthetic and real segmentations.

https://doi.org/10.1371/journal.pone.0333499.g006

To contextualize the segmentation results from synthetic 7T images, we first quantified the volumetric deviations between 3T-derived segmentations and those obtained from the 7T scans (Tables 6 and 7). This comparison is not intended to establish absolute accuracy as both 3T and 7T acquisitions are genuine ground-truth images, but rather to evaluate the degree of volumetric consistency each has with 7T-derived segmentations. Across the 21 clinically relevant regions examined, synthetic 7T volumes showed a deviation profile that was generally comparable to, and in several smaller subcortical structures closer to, the 7T-derived volumes than those obtained directly from 3T images. If, as suggested in prior literature, ultra-high-field 7T imaging provides more precise volumetric estimates due to its higher resolution and contrast, then these findings represent a clear support for the proposed synthesis method, demonstrating that synthetic 7T images can help bridge the gap between standard 3T acquisitions and the volumetric fidelity associated with 7T imaging.

thumbnail
Table 6. Comparison of ground truth 3T and 7T volumetric analysis.

https://doi.org/10.1371/journal.pone.0333499.t006

thumbnail
Table 7. Synthesis volume by brain region and difference between ground truth and synthetic sample.

https://doi.org/10.1371/journal.pone.0333499.t007

Table 7 shows the average volume across all 10-fold test volumes for both the synthetic and GT 7T MRI. VolBrain provides 132 different segmentation classes, but for this study we focus on a select 21 key structures of clinical relevance that we organize into 5 distinct groups. Table 2 from the methods section describes the clinical relevance of each of the structures, highlighting the potential implications and synthesis fidelity of certain clinically useful structures.

Table 8 shows the average asymmetry between select brain structures, including the hippocampus, caudate, thalamus, and lateral ventricles. The error between the ground truth and synthetic samples is also indicated.

thumbnail
Table 8. Synthesis volume by brain region and difference between ground truth and synthetic sample asymmetry indices.

https://doi.org/10.1371/journal.pone.0333499.t008

Finally, Fig 7 plots the segmentation mask overlayed on the ground truth and synthetic samples. The figure highlights the differences indicated by the structural data in Table 7, with the synthetic sample having increased white matter due to less clear boundaries between gray and white matter structures. Additionally, fine anatomical details and boundaries are smoothed out, as shown by the differences in brain cerebellum structural boundaries, shown by the pink structure.

thumbnail
Fig 7. Segmentation map comparison of subject 2 ground truth 7T (top) and synthetic 7T (bottom).

https://doi.org/10.1371/journal.pone.0333499.g007

Qualitative assessment

Finally, we present a qualitative analysis of the synthetic image regions. As shown in Fig 8, while the synthetic images faithfully capture the overall structure of gray and white matter, they struggle to preserve finer anatomical details with the same level of sharpness. High-contrast regions—particularly bright features absent in the original 3T scans, such as the area highlighted by the box—are consistently lost across all synthetic methods. However, compared to the original 3T input images, some anatomical regions show noticeable improvements in detail and resolution. Fig 9 illustrates a slice of the cerebellum in the original 3T, ground truth 7T, and synthetic MRI outputs. The areas highlighted by the blue boxes in the synthetic images exhibit enhanced anatomical detail compared to the 3T input, closely matching the 7T ground truth. This demonstrates how synthetic imagery can serve as a useful alternative high-field view for specific anatomical regions.

thumbnail
Fig 8. Comparison of subject 8 synthetic 7T and ground truth 7T and 3T T1W images.

https://doi.org/10.1371/journal.pone.0333499.g008

thumbnail
Fig 9. Cerebellum comparison of subject 8 synthetic 7T and ground truth 7T and 3T T1W images.

https://doi.org/10.1371/journal.pone.0333499.g009

The outputs from the FR-U-Net and semi supervised FR-U-Net models appear visually similar; however, the semi-supervised variant demonstrates improved sharpness and clarity in fine-detail regions. In contrast, the pix2pix baseline produces sharper-looking images overall, but introduces clear synthesis artifacts and omits important anatomical structures entirely.

3T and 7T acquisitions have distinct artifact profiles, with B1 field inhomogeneity being a known factor at 7T. In our dataset, such artifacts are present in the ground truth 7T images and therefore could be learned by the model during training. While our results suggest that the network preserved large-scale anatomical structures and contrast, it is possible that certain 7T-specific artifacts are also reproduced in the synthetic images. These qualitative patterns are consistent with our quantitative findings, where higher volumetric and boundary errors (Tables 6 and 7) occurred in small, high-curvature regions compared to larger cortical areas. At present, our method does not explicitly remove such artifacts. However, future work could incorporate artifact-robust training strategies, such as artifact simulation and augmentation, or adversarial discriminators trained to penalize known artifact patterns, to reduce their influence.

Discussion

This study presents a semi-supervised 3D multi-scale fusion residual U-Net (semi-supervised 3D FR-U-Net) for synthesizing high-resolution 7T MRI volumes from 3T inputs, with a novel evaluation framework centered on anatomical consistency. While the model demonstrated improved performance across standard synthesis metrics such as SSIM, PSNR, and NMSE, we argue that traditional pixel-level metrics are insufficient for determining real-world utility. To help bridge this gap, we introduced a segmentation-based evaluation pipeline using VolBrain, which allowed for interpretable, structure-specific analysis of anatomical fidelity across synthesized images.

Our model achieved the highest quantitative performance across all slice orientations compared to both fully supervised 3D U-Nets and a 2D slice-based pix2pix baseline. Most notably, it outperformed prior work by Siam et al. [18], who reported an SSIM of 71.07% and a PSNR of 18.62 using a CycleGAN-based method with data augmentation. Even in the transverse orientation, our model achieved a PSNR of 23.254 and SSIM of 73.682%. Although Siam et al. were able to increase SSIM above 80%, this required further preprocessing through skull stripping. Our methodology highlights the practicality of our approach in real-world clinical workflows, where extensive preprocessing may be undesirable or infeasible.

While our method demonstrates consistent improvements in SSIM, PSNR, and NMSE over existing baselines, we acknowledge that these gains are modest in absolute terms. This is not unexpected given the difficulty of synthesizing high-resolution, ultra-high-field MRI from a dataset of only 10 paired volumes, which is an extreme low-data regime by deep learning standards. In such settings, overfitting, noise amplification, and anatomical blurring are common challenges, and even marginal improvements can reflect meaningful architectural or learning-based advances. Moreover, traditional pixel-level metrics like SSIM and PSNR often undervalue perceptual and structural fidelity, particularly in clinically relevant regions. For example, sharper-looking images with slightly lower SSIM scores may yield more accurate anatomical segmentations or highlight diagnostically useful features. To address these limitations, our segmentation-based evaluation, focused on key substructures such as the hippocampus and cerebellum, offers a more informative view of region-specific strengths and weaknesses in anatomical preservation. Taken together, our results suggest that modest gains in conventional metrics, especially in low-data contexts, should be interpreted alongside structure-aware assessments that better reflect clinical relevance.

VolBrain analysis revealed that the model preserved major anatomical regions with high fidelity, achieving a Dice score of 0.9766 for whole-brain masks and greater than 0.84 for white and gray matter. Volumetric error was under 3% for most global and lobar structures, and asymmetry indices in regions such as the caudate and thalamus showed less than 6% deviation from ground truth. These results suggest the model could be clinically useful for tasks involving general brain morphology or basal ganglia analysis.

However, the same analysis also exposed important limitations. Despite relatively low volumetric error ( 2.9%) in the hippocampus, the asymmetry index deviated by more than 60%, indicating a substantial failure to preserve lateralized structural features. Given that inter-hemispheric hippocampal asymmetry is a critical marker in diagnosing Alzheimer’s disease and localizing seizure foci in epilepsy [7,37], this finding suggests that our model, in its current form, is not yet suitable for clinical applications involving the hippocampus or other fine subcortical structures. Hausdorff distance analysis further confirmed reduced boundary precision in small regions such as the amygdala and caudate compared to larger anatomical structures. Future studies should investigate fine-structure localized loss functions or the use of down-stream task guidance during training to help maintain fine grained anatomical fidelity.

These results reinforce the importance of the proposed segmentation-based evaluation framework. By moving beyond standard image similarity metrics and assessing anatomical fidelity in clinically meaningful regions, our approach can provide structural insight when radiologist review is infeasible—such as in large-scale studies or low-resource settings. While we did not have access to expert human annotators in this study, future work incorporating radiologist or neurologist input will be essential to define clinically acceptable thresholds for volumetric, boundary, and asymmetry errors. Such expert validation will be critical for aligning automated evaluation methods with diagnostic standards and improving the clinical readiness of synthesis models.

The study has several limitations. Most notably, it relies on a very small dataset of only 10 paired 3T–7T volumes, which limits generalizability across populations, acquisition protocols, and disease states. Although the semi-supervised framework allowed effective use of unpaired 7T data, larger and more diverse datasets will be necessary to assess robustness and performance under pathological conditions with abnormal structures. The network was trained on a fixed 3T–7T contrast mapping with specific acquisition parameters, which constrains its applicability across the wide range of MRI contrasts and protocols encountered in clinical practice. While the dataset also contains T2-weighted images, only T1-weighted images were used in this study to align with standard segmentation benchmarks such as those required by VolBrain and reduce modality-related variability. However, the architecture is contrast-agnostic and could be extended to T2-weighted and FLAIR images given appropriate paired datasets. These modalities provide complementary physiological information, such as sensitivity to edema, demyelination, and lesion load—not as prominent in T1-weighted imaging, and represent a promising direction for future work. This specificity reflects a broader limitation of current deep learning methods in MRI translation, where domain shift due to scanner settings, contrast weightings, and sequence variants can significantly impair performance.

Additionally, while we included a widely used baseline (pix2pix) and conducted detailed ablation studies to isolate the effects of key architectural components, comparisons to more recent state-of-the-art models (e.g., diffusion models, transformers) were not performed. This was due to either the lack of publicly available pretrained models for our problem setting or their incompatibility with small-volume paired datasets like ours. We acknowledge this as a limitation and have explicitly framed our evaluations in the context of low-data synthesis scenarios.

Future work should explore multi-modal synthesis to better reflect real-world neuroimaging protocols and develop models with contrast-adaptive or domain-generalizing capabilities. Creating large datasets with diverse acquisition parameters may greatly improve algorithm performance across real-world clinical imaging scenarios.

Future studies should also investigate region-specific learning strategies or model adaptations, such as incorporating transformer or diffusion-based modules tailored for low-data settings. Furthermore, integrating clinician-derived evaluation criteria—especially for high-impact regions like the hippocampus—will be essential for transitioning from research-grade synthesis to clinically actionable imaging. Incorporating segmentation-based loss guidance could further steer the network toward improved structural fidelity by explicitly reinforcing anatomical accuracy during training.

Conclusion

Overall, our semi-supervised 3D synthesis approach offers strong performance and promising anatomical fidelity in many brain regions, even under small data constraints. VolBrain-based segmentation revealed that the model preserves overall brain structure and may support applications in certain contexts, such as evaluating caudate volume or global brain atrophy. However, significant limitations remain in regions like the hippocampus, underscoring the need for targeted improvements and radiologist-informed evaluation frameworks. Additionally, the current results must be considered in the context of potential domain shifts, scanner- and protocol-specific dependencies, and the lack of validation on pathological cases, any of which could limit their generalizability to broader clinical populations. These findings mark an important step toward clinically interpretable synthetic MRI and demonstrate how segmentation-based assessments can clarify the true readiness of generative models for real-world use.

Acknowledgments

The authors thank the creators of the public datasets used in this study, which are cited in the manuscript.

References

  1. 1. Plewes DB, Kucharczyk W. Physics of MRI: A primer. J Magn Reson Imaging. 2012;35(5):1038–54. pmid:22499279
  2. 2. Deoni SCL, Meyers SM, Kolind SH. Modern methods for accurate T1, T2, and proton density MRI. Oxford textbook of neuroimaging. Oxford University Press; 2015. p. 13–26. https://doi.org/10.1093/med/9780199664092.003.0002
  3. 3. Chavhan GB, Babyn PS, Thomas B, Shroff MM, Haacke EM. Principles, techniques, and applications of T2*-based MR imaging and its special applications. Radiographics. 2009;29(5):1433–49. pmid:19755604
  4. 4. Khoo MMY, Tyler PA, Saifuddin A, Padhani AR. Diffusion-weighted imaging (DWI) in musculoskeletal MRI: A critical review. Skeletal Radiol. 2011;40(6):665–81. pmid:21311884
  5. 5. Zwanenburg JJM, Hendrikse J, Visser F, Takahara T, Luijten PR. Fluid attenuated inversion recovery (FLAIR) MRI at 7.0 Tesla: Comparison with 1.5 and 3.0 Tesla. Eur Radiol. 2010;20(4):915–22. pmid:19802613
  6. 6. Okada T, Fujimoto K, Fushimi Y, Akasaka T, Thuy DHD, Shima A, et al. Neuroimaging at 7 Tesla: A pictorial narrative review. Quant Imaging Med Surg. 2022;12(6):3406–35. pmid:35655840
  7. 7. Suthana NA, Donix M, Wozny DR, Bazih A, Jones M, Heidemann RM, et al. High-resolution 7T fMRI of human hippocampal subfields during associative learning. J Cogn Neurosci. 2015;27(6):1194–206. pmid:25514656
  8. 8. Cosottini M, Roccatagliata L. Neuroimaging at 7 T: Are we ready for clinical transition?. Eur Radiol Exp. 2021;5(1):37. pmid:34435257
  9. 9. Dayarathna S, Islam KT, Uribe S, Yang G, Hayat M, Chen Z. Deep learning based synthesis of MRI, CT and PET: Review and analysis. Med Image Anal. 2024;92:103046. pmid:38052145
  10. 10. Osman AFI, Tamam NM. Deep learning-based convolutional neural network for intramodality brain MRI synthesis. J Appl Clin Med Phys. 2022;23(4):e13530. pmid:35044073
  11. 11. Zhao P, Pan H, Xia S. MRI-Trans-GAN: 3D MRI cross-modality translation. In: 2021 40th Chinese control conference (CCC); 2021. https://doi.org/10.23919/ccc52363.2021.9550256
  12. 12. Uzunova H, Ehrhardt J, Handels H. Memory-efficient GAN-based domain translation of high resolution 3D medical images. Comput Med Imaging Graph. 2020;86:101801. pmid:33130418
  13. 13. Wei J, Pan Y, Xia Y, Shen D. Learning to synthesize 7 T MRI from 3 T MRI with few data by deformable augmentation. In: Lian C, Cao X, Rekik I, Xu X, Yan P, editors. Machine learning in medical imaging. Cham: Springer International Publishing; 2021. p. 70–9.
  14. 14. Cui Q, Tosun D, Mukherjee P, Abbasi-Asl R. 7T MRI synthesization from 3T acquisitions. In: Linguraru MG, Dou Q, Feragen A, Giannarou S, Glocker B, Lekadir K, et al., editors. Medical image computing and computer assisted intervention – MICCAI 2024 . Cham: Springer Nature Switzerland; 2024. p. 35–44.
  15. 15. Chen X, Qu L, Xie Y, Ahmad S, Yap P-T. A paired dataset of T1- and T2-weighted MRI at 3 Tesla and 7 Tesla. Sci Data. 2023;10(1):489. pmid:37500686
  16. 16. Yurt M, Dalmaz O, Dar S, Ozbey M, Tinaz B, Oguz K, et al. Semi-supervised learning of MRI synthesis without fully-sampled ground truths. IEEE Trans Med Imaging. 2022;41(12):3895–906. pmid:35969576
  17. 17. Liu X, Qiu S, Zhou M, Le W, Li Q, Wang Y. WFANet-DDCL: Wavelet-based frequency attention network and dual domain consistency learning for 7T MRI synthesis From 3T MRI. IEEE Trans Circuits Syst Video Technol. 2025;35(6):5617–32.
  18. 18. Shams Siam Z, Tasnuva Hasan R, Hossain Chowdhury M, Shaheenur Islam Sumon Md, Bin Ibne Reaz M, Hamid Bin Md Ali S, et al. Improving MRI resolution: A cycle consistent generative adversarial network-based approach for 3T to 7T translation. IEEE Access. 2024;12:116498–515.
  19. 19. Coupé P, Mansencal B, Clément M, Giraud R, Denis de Senneville B, Ta V-T, et al. AssemblyNet: A large ensemble of CNNs for 3D whole brain MRI segmentation. Neuroimage. 2020;219:117026. pmid:32522665
  20. 20. Bahrami K, Shi F, Rekik I, Gao Y, Shen D. 7T-guided super-resolution of 3T MRI. Med Phys. 2017;44(5):1661–77. pmid:28177548
  21. 21. Bahrami K, Shi F, Zong X, Shin HW, An H, Shen D. Reconstruction of 7T-like images from 3T MRI. IEEE Trans Med Imaging. 2016;35(9):2085–97.
  22. 22. Bahrami K, Shi F, Rekik I, Shen D. Convolutional neural network for reconstruction of 7T-like images from 3T MRI using appearance and anatomical features. In: Carneiro G, Mateus D, Peter L, Bradley A, Tavares JMRS, Belagiannis V, et al., editors. Deep learning and data labeling for medical applications. Cham: Springer International Publishing; 2016. p. 39–47.
  23. 23. Wu Z, Chen X, Xie S, Shen J, Zeng Y. Super-resolution of brain MRI images based on denoising diffusion probabilistic model. Biomed Signal Process Control. 2023;85:104901.
  24. 24. Duan C, Bian X, Cheng K, Lyu J, Xiong Y, Xiao S, et al. Synthesized 7T MPRAGE from 3T MPRAGE using generative adversarial network and validation in clinical brain imaging: A feasibility study. J Magn Reson Imaging. 2024;59(5):1620–9. pmid:37559435
  25. 25. Qu L, Zhang Y, Wang S, Yap P-T, Shen D. Synthesized 7T MRI from 3T MRI via deep learning in spatial and wavelet domains. Med Image Anal. 2020;62:101663. pmid:32120269
  26. 26. Lin H, Figini M, Tanno R, Blumberg SB, Kaden E, Ogbole G, et al. Deep learning for low-field to high-field MR: Image quality transfer with probabilistic decimation simulator. In: Knoll F, Maier A, Rueckert D, Ye JC, editors. Machine learning for medical image reconstruction. Cham: Springer International Publishing; 2019. p. 58–70.
  27. 27. Figini M, Lin H, Ogbole G, Arco FD, Blumberg SB, Carmichael DW, et al. Image quality transfer enhances contrast and resolution of low-field brain MRI in African paediatric epilepsy patients; 2020. Available from: http://arxiv.org/abs/2003.07216.
  28. 28. Li X, Wang L, Liu H, Ma B, Chu L, Dong X, et al. Syn_SegNet: A joint deep neural network for ultrahigh-field 7T MRI synthesis and hippocampal subfield segmentation in routine 3T MRI. IEEE J Biomed Health Inform. 2023;27(10):4866–77. pmid:37581964
  29. 29. Eidex Z, Safari M, Wang T, Wildman V, Yu DS, Mao H, et al. Generalizable 7T T1-map synthesis from 1.5T and 3T T1 MRI with an efficient transformer model; 2025. http://arxiv.org/abs/2507.08655
  30. 30. Zhang Y, Cheng JZ, Xiang L, Yap PT, Shen D. Dual-domain cascaded regression for synthesizing 7T from 3T MRI. In: Frangi AF, Schnabel JA, Davatzikos C, Alberola-López C, Fichtinger G, editors. Medical image computing and computer assisted intervention – MICCAI 2018 . Cham: Springer International Publishing; 2018. p. 410–7.
  31. 31. Alom MZ, Hasan M, Yakopcic C, Taha TM, Asari VK. Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation; 2018. http://arxiv.org/abs/1802.06955
  32. 32. Isola P, Zhu JY, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks; 2018. http://arxiv.org/abs/1611.07004
  33. 33. Despotovi ć I, Goossens B, Philips W. MRI segmentation of the human brain: Challenges, methods, and applications. Comput Math Methods Med. 2015;2015:450341. pmid:25945121
  34. 34. Fawzi A, Achuthan A, Belaton B. Brain image segmentation in recent years: A narrative review. Brain Sci. 2021;11(8):1055. pmid:34439674
  35. 35. Singh MK, Singh KK. A review of publicly available automatic brain segmentation methodologies, machine learning models, recent advancements, and their comparison. Ann Neurosci. 2021;28(1–2):82–93. pmid:34733059
  36. 36. Tang X, Crocetti D, Kutten K, Ceritoglu C, Albert MS, Mori S, et al. Segmentation of brain magnetic resonance images based on multi-atlas likelihood fusion: Testing using data with a broad range of anatomical and photometric profiles. Front Neurosci. 2015;9:61. pmid:25784852
  37. 37. Pusparani Y, Lin C-Y, Jan Y-K, Lin F-Y, Liau B-Y, Alex JSR, et al. Hippocampal volume asymmetry in Alzheimer disease: A systematic review and meta-analysis. Medicine (Baltimore). 2025;104(10):e41662. pmid:40068028