Figures
Abstract
Pixel-level annotation of lung cavities (LCs) in computed tomography (CT) images is challenging due to their morphological diversity and complexity. Weakly supervised semantic segmentation (WSSS) methods, which utilize sparse annotations (e.g., image-level labels), offer a promising solution. However, existing WSSS approaches often generate coarse pseudo-labels and lack sufficient spatial supervision, resulting in under- or over-segmentation of irregular lesions. To address these limitations, we introduce several key innovations. First, we propose a novel Graph-based Affinity Network (GA-Net) that, unlike conventional methods relying on low-level pixel features, models long-range contextual relationships and structural dependencies using a superpixel graph and learned edge inference kernel, enabling structure-aware pseudo-label refinement for complex lesion morphology. Second, we introduce region-wise affinity propagation, which refines segmentation by propagating activations within semantically coherent 3D regions, offering more precise control over under-/over-segmentation compared to global affinity methods. Additionally, we incorporate Exponential Moving Average (EMA) ensembling for training stability and a scribble-based segmentation module that utilizes pseudo-label contours to provide direct boundary supervision. Extensive experiments on three benchmark datasets demonstrate that our method outperforms existing state-of-the-art medical WSSS techniques, achieving precise and reliable segmentation of complex LCs in CT scans.
Citation: Ding Z, Tan Z, Madzin H, Li Z, Liu J (2026) Refining weak supervision for robust lung cavity segmentation: A graph-affinity method with boundary constraints. PLoS One 21(2): e0341717. https://doi.org/10.1371/journal.pone.0341717
Editor: Hongchuan Yu, Bournemouth University, UNITED KINGDOM OF GREAT BRITAIN AND NORTHERN IRELAND
Received: August 17, 2025; Accepted: January 12, 2026; Published: February 10, 2026
Copyright: © 2026 Ding et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The minimal dataset underlying the conclusions of this manuscript is available in Figshare at https://doi.org/10.6084/m9.figshare.31033282.
Funding: Research Management Centre UPM and Faculty of Computer Science and Information Technology, Universiti Putra Malaysia for Journal Publication Grant.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Lung cavities (LCs) are crucial radiographic indicators for diagnosing tuberculosis (TB), playing an essential role in confirming diagnoses, assessing disease progression, and monitoring the efficacy of treatment regimens [1,2]. However, due to the variability in the shape and size of LCs, accurately annotating these LCs at the pixel-level in computed tomography (CT) scan is highly challenging [2]. In recent years, deep learning (DL) has experienced rapid advancements, leading to its widespread adoption in medical image analysis [3–6], particularly in the identification of pulmonary TB LCs [2,7]. In addition, due to the data-hungry nature of DL technology, segmentation models based on fully supervised training paradigms typically require vast amounts of labor-intensive accurate LCs pixel-level annotation data for learning [3–6]. Obtaining these precise pixel-level annotations for LCs often requires a significant amount of human and material resources. Utilizing sparse labels in weakly supervised semantic segmentation (WSSS) methods has become an important trend to overcome this limitation [2,8–10]. These methods train semantic segmentation models using image-level [2,10], scribble [11,12], or bounding box [13] annotations as supervision signals.
Although some weakly supervised segmentation methods perform well in other imaging domains [14–16], they often lack versatility when directly applied to medical lesion tissue segmentation (e.g., TB LCs recognition), making it difficult to achieve good recognition results [2,12,17–19]. In particular, two major challenges often arise when applying existing weakly supervised segmentation methods specifically to TB LCs segmentation research: 1) In the pseudo-label generation stage, many conventional methods [20–24] rely on initial seeds derived from Class Activation Mapping (CAM) [25] and refine them through affinity propagation [21,22,24], such as the random walk [26] or bilateral affinity models adopted in works like IRNet [21]. However, the limitations of these conventional affinity propagation algorithms become pronounced when dealing with LCs, which often exhibit irregular shapes, thin walls, or internal septations. These methods typically construct affinity matrices based on local, low-level pixel appearance similarities (e.g., color and texture), lacking effective modeling of long-range contextual relationships and complex structural dependencies [20–24]. As a result, the propagation process is prone to failure at cavity wall discontinuities or in low-contrast regions, leading to persistent under-segmentation (failure to cover the entire cavity region) or over-segmentation (incorrect inclusion of adjacent normal tissues) in the generated pseudo-labels, as illustrated in Fig 1. 2) At the level of supervision signals, WSSS paradigms that rely exclusively on image-level labels–such as classical CAM-based methods and their variants–provide extremely sparse and weak supervision to the model. Under this setting, models tend to focus on the most discriminative local regions (e.g., the most visually salient parts of a cavity) while failing to perceive and learn the complete spatial extent, precise morphology, and detailed boundary characteristics of the target. Although subsequent studies [27,28] have attempted to alleviate this issue by incorporating self-attention mechanisms or boundary-aware constraints, these approaches may struggle to obtain reliable learning signals when the pseudo-label quality is poor during early training stages, thereby limiting their performance ceiling in complex medical image segmentation tasks.
A: conventional affinity method. B: graph-based affinity methods.
In this paper, we propose a novel WSSS method that introduces targeted innovations to directly address these two core challenges, as shown in Fig 2. Our primary contribution is a new affinity learning paradigm for pseudo-label refinement. Unlike previous methods [20–24], we propose a Graph-based Affinity Network (GA-Net) that constructs a superpixel graph to represent image structure and employs a trainable network as a kernel to infer affinities, enabling structure-aware, long-range diffusion tailored for complex lesions. Complementing this, we design a Region-wise Affinity Propagation mechanism that refines the CAM by propagating activation scores from high-confidence regions to semantically similar, low-activation regions. This mechanism effectively addresses the problem of coarse pseudo-labels generated by CAM, enhancing the localization of LCs and overcoming the limitations of under-segmentation or over-segmentation. By using a learned affinity matrix, our method leverages regional information to propagate affinities within smaller 3D regions, leading to a more precise segmentation. Additionally, to enhance the robustness of our model, we integrate Exponential Moving Average (EMA) ensembling, which stabilizes model predictions in the early stages of training, reducing the impact of transient errors. This technique smooths the predictions, allowing for better consistency in the refinement of pseudo-labels during training. To further refine the segmentation of small and challenging lesions, we incorporate a scribble-based segmentation module. This module utilizes sparse scribble annotations generated from pseudo-label contours, enhancing the model’s ability to capture fine boundary details of small lesions. A partial cross-entropy loss is used to focus the model’s attention on the annotated lesion areas, improving its performance in detecting and segmenting small lesions that might otherwise be overlooked. Through extensive experiments on three standard datasets, our proposed method demonstrates superior performance compared to existing state-of-the-art 3D WSSS methods.
Methodology
Intuitive overview. The core objective of GA-Net is to integrate appearance features, spatial context, and the weakly-supervised localization cues provided by the CAM to infer whether two superpixel regions belong to the same anatomical structure. To avoid the high computational cost and noise sensitivity of pixel-wise affinity propagation, the proposed method first partitions the image into superpixels and constructs a compact region-adjacency graph on top of them. In this graph, each node corresponds to a superpixel region, and edges connect region pairs that are either spatially adjacent or feature-similar, thereby encoding potential semantic relationships between them. GA-Net learns the affinity weights of the graph edges in an end-to-end, data-driven manner. The supervision signal originates from the soft overlap measure between regions derived from the CAM. Based on this, the network predicts a soft affinity score that quantifies the probability of region co-membership. The learned affinity matrix is then used to construct a random-walk transition matrix, which propagates and diffuses the initial CAM confidences. This process spreads high-confidence information along anatomically consistent paths to semantically related regions while suppressing spurious connections. The entire pipeline is fully differentiable, allowing GA-Net to be integrated with the backbone segmentation network for end-to-end joint training, as shown in Algorithm 1.
Algorithm 1 End-to-end training and inference pipeline of GA-Net.
Require: Input image I, image-level label b
Ensure: Refined segmentation mask
1: 1. Feature extraction and graph construction
2: Extract encoder features
3: Generate superpixel regions {P(i)} using SLIC
4: Construct graph , including adjacency edges and kNN long-range edges
5: Initialize node features:
6: Generate soft target affinity w from CAM (Equation 2)
7: 2. GA-Net inference for edge affinity
8: for to L–1 do
9: for each edge do
10: Compute attention weight:
11: end for
12: for each node do
13: Update node feature:
14: end for
15: end for
16: Compute edge posterior probability:
17: 3. Loss computation and optimization
18: Compute structural consistency loss:
19: Compute smoothing regularization term:
20: Total loss:
21: Backpropagate and update GA-Net parameters
22: 4. Random-walk propagation (inference phase)
23: Construct transition matrix: , where
24: Propagate activations:
25: return
Graph-based affinity learning
We formulate affinity refinement as probabilistic edge inference on a superpixel graph, where the goal is to estimate the posterior probability that two regions belong to the same anatomical structure. A GA-Net serves as a learned kernel producing edge posteriors under weak supervision from CAM, while random-walk propagation realizes structure-aware diffusion over a row-stochastic transition matrix. Below, we detail the derivations and design choices.
Graph construction as a latent-variable model: For any pair of superpixels connected in the graph, we define a binary latent variable , where y = 1 indicates that the two regions belong to the same anatomical structure, and y = 0 otherwise. Our goal is to estimate the posterior probability of this co-membership:
where I denotes the input image or volume, y represents region co-membership, and is the true posterior affinity between the two connected regions. Since direct supervision for y is unavailable, a soft target is instead derived from the CAM overlap:
where P(1) and P(2) are the pixel/voxel sets corresponding to the two regions, and w represents the soft target affinity based on the overlap between these regions in the CAM. This soft target captures the expected co-membership under weak supervision. To reduce the dimensionality of the latent space, we apply SLIC (Simple Linear Iterative Clustering) [29] segmentation, which converts the pixel-level graph into a compact region graph , where the feature vector for each node is initialized as:
where represents the image encoder feature,
is the intensity appearance feature,
is the spatial coordinate encoding (e.g., centroid), and
is the initial affinity. The operator
denotes concatenation of these features.
k-Nearest Neighbor (kNN) long-range edges: Anatomical structures, such as elongated vessels, may be spatially separated but semantically similar. To account for this, we incorporate kNN edges in the feature space. This reduces the graph diameter and enables multi-hop consistency, while maintaining local planarity with adjacency edges. These long-range edges help propagate affinity information across regions that are anatomically related but spatially distant.
GA-Net as a learned positive-definite Kernel: A single GA-Net layer computes the normalized attention weights using a learned similarity measure:
where is the learnable projection matrix,
is the attention vector,
and
are the features of the two connected nodes,
represents the normalized attention weight,
is the neighborhood set, and
is the activation function. The nodes indexed by (1) and (2) denote the two nodes in the edge, and no specific spatial dimensionality is implied.
Attention vs. Fixed Kernels: Medical textures in images are inherently non-stationary. By using learned attention, the model can adapt to the specific context of the image. This allows attention to down-weight misleading cues (such as noise or specularities) and up-weight important features like those from self-attention network near anatomical boundaries. Furthermore, the concatenation-based scoring method allows the model to capture cross-feature interactions that static kernels.
From node embeddings to edge posteriors: For each edge, the Bernoulli parameter [30] is predicted as:
This is equivalent to performing logistic regression on the final GA-Net embeddings, transforming the problem into a maximum-likelihood estimation task under the weak labels w.
Loss as Maximum-A Posteriori (MAP) estimation: The overall loss function combines a penalized negative log-likelihood term:
Structural consistency (data term): Since w serves as a soft target for y, we use soft-label cross-entropy as the loss function:
This is equivalent to minimizing the Kullback-Leibler divergence between the Bernoulli distributions and
. The inclusion of both positive and negative terms in the loss helps reduce false-positive affinities, particularly in noisy CAM scenarios. While the original formulation approximates a positive-class upper bound, the full soft binary cross-entropy (BCE) loss is more robust and less prone to overfitting. A practical note is that binary CAM can be replaced with a soft Dice/Jaccard form to ensure differentiability:
Local smoothness (informative prior): To stabilize the model, a Tikhonov prior [31] is applied to the initial affinity:
Here, is a confidence weight based on feature similarity or edge strength. This prior helps to anchor the weak supervision around the initial structure and accelerates convergence, while also preventing overly sharp or discontinuous affinity distributions. From a MAP perspective, it models the posterior as:
This smoothness prior is critical for ensuring numerical stability, particularly in random walk-based propagation, and contributes to the overall robustness of the model.
Class-driven affinity pseudo label generation
Let denote the classification head, which comprises fully connected layers. We can derive the initial pseudo-label Yp for an input medical CT image
labeled with binary categories
as:
, where every pixel contributes to the generation of the pseudo-label. Specifically, to generate the initial pseudo-label,
is trained using image-level labels with label smoothing cross-entropy loss [32], formulated as:
where . K represents the number of classes. This adjustment ensures a more balanced learning process by softening the targets, thereby improving the model’s robustness against noisy labels.
Initial pseudo label generation: Deriving a reliable initial pseudo label Yaff is a crucial step in learning the semantic affinity Waff. As illustrated in Fig 2, we extract Yaff from initial pseudo labels Yp. Following [2,14,26], we utilize two thresholds, and
, where
, to categorize the Yp into three distinct regions: reliable foreground, background, and uncertain areas. Formally, given the CAM
, the pseudo label Yp is constructed as follows:
where 0 denotes the background and 1 denotes ignored regions. The function identifies the semantic class with the highest activation value.
Affinity Converter (AC): The Affinity Converter (AC) serves as a crucial component in our method, transforming raw feature representations into meaningful affinity relationships. For affinity elements in the three-dimensional affinity matrix, it represents the affinity between the voxel at
and
. The calculation of affinity can similarly be based on the vector representation of voxels
, as shown:
Focal affinity loss: The generated Yaff will be used as a supervisory signal to guide the model in learning to predict an accurate semantic affinity representation Waff. This supervision process is achieved by minimizing the discrepancy between the model’s prediction Waff and the ground-truth label Yaff. Specifically, we use a modified focal affinity loss term , which takes the following form:
where N+ and N−- denote the counts of and
samples.
and
represent the sets of positive and negative samples in Yaff, respectively.
Region-wise affinity propagation: To further refine the CAM using learned affinities, we introduce a Region-wise Affinity Propagation mechanism. This mechanism enables the propagation of activation scores from high-confidence regions to semantically similar but low-activation regions, leveraging the learned affinity matrix. Let the CAM volume be , and let the global semantic affinity matrix be
. We partition the CAM into R overlapping local 3D regions, where each region
contains a subvolume M
and corresponding affinity submatrix
. To perform affinity-based propagation within each region, we first construct a row-normalized transition matrix
based on the affinity submatrix:
where controls the propagation strength, and Dr is the diagonal normalization matrix to ensure each row of Tr sums to 1. The refined activation
for region r is computed by multiplying the transition matrix with the vectorized CAM:
To aggregate results across overlapping regions, we apply a distance-weighted merging strategy. Let denote all regions containing voxel (i, j, k), and let wr(i, j, k) be the weight assigned to region r based on the voxel’s proximity to the region center cr. The final propagated CAM is calculated as:
Basic supervised semantic segmentation
The initial pseudo labels Yp undergo a series of refinement processes, ultimately generating the pseudo labels for the semantic segmentation task. In the pseudo-annotated semantic segmentation training, we adopt a smoothed cross-entropy loss
:
where is a regularization term for label smoothing [32] to prevent the model from becoming overconfident. K is the number of classes.
is the prediction from the model, yi is the one-hot label.
Exponential moving average: In the initial stages of training for WSSS, model predictions are often unstable and exhibit limited accuracy. To address this challenge, we propose an approach inspired by [11], wherein we employ a prediction ensembling technique known as Exponential Moving Average (EMA). This technique continuously accumulates and updates the EMA of predictions during training, effectively reducing the adverse impact of single prediction errors on the final results. Specifically, at each training iteration, the EMA combines the current model prediction with the previous EMA prediction yn−1, where the weights are determined by the smoothing factor δ, following the update rule:
where yn represents the current EMA prediction, is the current model prediction, yn−1 is the previous EMA prediction, and
is the smoothing factor. The predictions are averaged every
iterations, where
is the ensembling interval.
Scribbled-based segmentation
To accurately segment small LC lesions, we propose a scribble-based segmentation approach. This method leverages scribble annotations to capture the complete shape and boundary information of LC target objects, thereby enhancing the model’s ability to identify and process fine structures [12]. Unlike expert manual labels, these scribble annotations are indirectly generated by extracting and converting the contours of the pseudo labels using the Lung Cavity Recognition Algorithm (LCRA) described in [2]. The pseudo labels provide approximate boundaries, which are then transformed into scribble annotations for improved LC segmentation accuracy. Specifically, we use a partial cross-entropy loss function
as a sparse annotation supervision paradigm. This loss is defined as follows:
where I is the input CT image, is the scribble label derived from the pseudo-label Ω, and
is the set of scribbled pixels.
represents the model’s prediction.
Data and experiments
Dataset
We conducted experiments on three publicly available TB chest CT datasets, covering both classification and segmentation tasks. Details of the datasets and data splitting strategies are summarized as follows. The classification criteria for lung cavity number and size across the three datasets are summarized in Table 1.
ImageCLEF Tuberculosis (https://www.imageclef.org/2021/medical/tuberculosis): This dataset [33] contains chest CT scans of 1,338 TB patients, each corresponding to a single patient and labeled with only one TB type. Some scans are accompanied by additional meta-information, which may vary by case. Among them, 917 images are designated for training and 421 for testing.
TB Portals (https://tbportals.niaid.nih.gov/): A global TB database and research platform developed by the National Institutes of Health (NIH) [34], containing 1,324 non-compound TB CT scans along with corresponding clinical data. We used the same categorical variables (LCs size and total cavity number), as shown in Table 1. In our study, 1,059 images were used for training and 265 for validation.
DeepPulmoTB (https://github.com/SupCodeTech/DeepPulmoTB): This dataset [7] includes 354 CT images with multi-class annotations for TB, including lung areas, LCs, and consolidation or lung cavity wall (C-LCW). Due to the limited dataset size, we adopted a 9:1 split strategy, using 318 images for training and 36 for validation.
Label generation and refinement
Image-level annotation generation. Notably, in the Table 1, the ImageCLEF TB dataset does not provide annotations specific to LCs classification or segmentation, and the TB Portals dataset lacks pixel-level segmentation labels. To address this limitation, we collaborated with experts from the Department of Radiology at Universiti Putra Malaysia to construct reliable image-level supervision. The annotation process was conducted manually by three senior radiologists, each with over ten years of clinical experience in thoracic imaging. All annotations were performed using ITK-SNAP [35] under a unified and standardized protocol agreed upon by all annotators. Each patient CT volume was assigned image-level labels (e.g., 1-3 cavities and <25mm), which describe the shape characteristics of LC lesions and serve as the sole human supervision during the initial classification and CAM generation stages.
Pseudo-label derivation and refinement. Based on the image-level labels, a classification head was first trained to produce CAMs. These CAMs provide coarse localization cues but are inherently noisy and incomplete. To obtain initial pseudo pixel-level labels Yp, we applied a dual-threshold strategy with thresholds and
, partitioning CAM responses into reliable foreground, background, and uncertain regions. Subsequently, structural affinity learning and region-wise affinity propagation were employed to refine these pseudo labels, resulting in progressively improved pseudo segmentation masks
.
Scribble annotation generation and supervision. Unlike conventional approaches that rely on manually annotated scribbles, the scribble annotations in our method are generated automatically. Specifically, we extract the contours of the refined pseudo labels and convert them into sparse scribble annotations using the Lung Cavity Recognition Algorithm (LCRA) proposed in [2]. These scribbles preserve essential shape and boundary information of lung cavity lesions while avoiding the need for additional expert annotation. The resulting scribble masks are then used to construct partial supervision masks
, where only scribbled pixels contribute to the loss computation via a partial cross-entropy objective. This strategy enables effective learning from sparse supervision and further improves segmentation accuracy for small and irregular LCs lesions.
Overall, the proposed pipeline establishes a complete and coherent supervision hierarchy, starting from expert-provided image-level labels, progressing through affinity-refined pseudo-labels, and finally yielding scribble-based supervision masks for robust weakly supervised semantic segmentation.
Data preprocessing and experimental setup
To ensure data consistency and model training stability, we applied a unified preprocessing pipeline to all datasets.
Imaging modality and specifications: All data were obtained from chest CT volumes acquired using standard clinical scanning protocols, in the original DICOM format. The specific parameters of CT images varied among cases: the in-plane resolution ranged from mm to
mm, the slice thickness ranged from 1 mm to 5 mm, and each 3D volume contained approximately 100 to 400 consecutive slices. To eliminate spatial scale differences among data from different sources, we resampled all CT volumes to a uniform isotropic resolution of
mm3. Furthermore, the image gray values (Hounsfield Units, HU) were clipped to the window of [–1200,600] to focus on the typical density range of lung tissue, and then linearly normalized to the interval [0,1] for model input.
Data augmentation: During the training phase, we adopted an online data augmentation strategy based on MONAI to improve the generalization ability of the model. For each CT volume, we first loaded the image and its corresponding segmentation label, and uniformly converted the data into the channel-first format (). Then, fixed-size three-dimensional patches were randomly cropped from the original volume, with the patch size set to
and random size variation disabled (random_size=False). This random 3D cropping was performed online at each iteration to increase the spatial diversity of training samples and reduce memory overhead. For intensity augmentation, we applied random Gaussian noise perturbation (RandGaussianNoise/RandGaussianNoised) only to the input images (img) to simulate imaging noise and intensity fluctuations caused by different scanning conditions, thereby enhancing the model’s robustness to noise and intensity variations. It is important to emphasize that all spatial transformations (e.g., random cropping) were applied synchronously to both images and segmentation labels to ensure pixel-level alignment, while intensity-based transformations were applied only to images and not to segmentation labels, so as to avoid introducing label noise.
Hardware and software environment: In our experimental setup, all models were trained on a single NVIDIA RTX 3090Ti GPU with 24 GB of memory. During training, the batch size was set to 8, and the initial learning rate was . Model optimization was performed using the AdamW optimizer, with a weight decay of 0.05 and momentum parameters
and
. A warm-up strategy was applied at the beginning of training, during which the learning rate was linearly increased to 5.625 × 10−5, followed by a cosine annealing learning rate scheduler to gradually decay the learning rate.
Cross-validation protocol and statistical evaluation
To clarify the robustness of the statistical analysis and the evaluation protocol, we provide additional details on the cross-validation (CV) design and inferential testing strategy adopted in this study. For each dataset, a fixed held-out test set was first defined according to the protocol described in the Dataset section. The remaining data were used exclusively for training and validation. Within this training set, a 5-fold cross-validation strategy was employed for hyper-parameter selection and model configuration. Specifically, the training data were randomly partitioned into five non-overlapping folds at the patient level. In each CV iteration, four folds were used for model training, and the remaining fold was used for validation. This process was repeated five times, such that each fold served as the validation set once. The patient-level data distribution for the 5-fold cross-validation across all datasets is summarized in Table 2. All hyperparameters, including those related to graph-based affinity learning, pseudo-label generation, and supervised segmentation, were selected based solely on the average validation performance across the five folds. Importantly, the held-out test set was never used during cross-validation, model selection, or hyper-parameter tuning. After determining the optimal configuration, the final model was retrained using the full training set and evaluated once on the held-out test set.
For each dataset, a fixed test set is held out and excluded from cross-validation. Five-fold CV is performed only on the training pool at the patient level.
Moreover, to ensure robust and statistically sound comparisons, significance testing was performed on segmentation metrics (DSC, IoU) rather than on correlated training losses. Specifically, for each method, we obtained its prediction for every individual patient in the dataset. Since all methods were evaluated on the same set of patients, the DSC/IoU scores across patients formed paired samples for any pair of methods. We then applied the Wilcoxon signed-rank test to these paired patient-level scores to determine statistical significance. This approach provides a more reliable and powerful inference than using fold-aggregated scores (which would yield only 5 data points per method) or training losses. Table 3 summarizes the statistical evaluation protocol. All statistical analyses were performed using Python’s SciPy library (version 1.10.1)
Implementation details
To ensure reliable evaluation and stable model training, all experiments were conducted using a consistent experimental setup. The image encoder was fixed to ViT-B/16 as the backbone network across all experiments. Details of the cross-validation protocol, including data partitioning, hyper-parameter selection, and statistical evaluation, are provided in the previous section.
Sensitivity analysis of key hyper-parameters for graph-based affinity learning. To assess the robustness of the proposed method with respect to key hyper-parameters, we conducted a sensitivity analysis by varying the affinity thresholds, graph design parameters, and propagation settings around their default values. Specifically, the foreground and background thresholds were perturbed within a reasonable range (e.g.,
% of the default values), and the radius of the local 3D window was varied between (5,5,5) and (9,9,9). In addition, we evaluated alternative GA-Net depths in the range of 1 to 15 layers.
1) Pseudo-label Generation Thresholds : These thresholds are used to partition reliable foreground, background, and uncertain regions from the initial CAM, and play a critical role in affinity learning. To evaluate the sensitivity of the proposed method to these parameters, we systematically examined different combinations of
and
within the ranges [0.25,0.35] and [0.40,0.50], respectively. Experimental results show that when
varies within these ranges, the final segmentation performance, measured by the DSC, remains on a consistently high and stable plateau with only minor fluctuations. The selected configuration
lies well within this high-performance region. Further statistical testing (patient-level paired Wilcoxon signed-rank test) confirms that there is no statistically significant difference in performance between the selected configuration
and other combinations within the interval (all comparisons p > 0.05). This indicates that the method is robust to small perturbations in the threshold parameters, and its effectiveness stems from the overall design rather than from precise tuning of specific threshold values.
2) Local Affinity Window Radius: Computing affinities within a local 3D window balances capturing sufficient context with computational efficiency and noise suppression. We tested cubic window sizes from 5 to 11 voxels per side. Performance peaked at a radius of 7 (i.e., a (7,7,7) window). Importantly, the performance degradation observed at radii of 5 or 9 compared to the peak performance was statistically non-significant (p > 0.05, patient-level paired Wilcoxon test). This suggests that the method does not critically depend on an exact window size and can achieve stable performance as long as it operates within a reasonable range of local context.
3) GA-Net Depth on Graph-Based Affinity Learning: For the graph-based affinity learning module, we systematically investigated the effect of GA-Net depth by varying the number of layers from 1 to 15. The DSC increased steadily as the network depth grew from 1 to 4 layers, indicating that deeper message passing enables more effective aggregation of contextual and structural information across superpixel regions. The performance reached its peak at four layers (as illustrated in Fig 3(a)), suggesting a favorable balance between local feature preservation and contextual integration. However, when the number of layers exceeds five, the segmentation performance exhibits a statistically significant decline (e.g., comparing 6 layers to 4 layers, p < 0.05). This performance degradation primarily stems from two aspects: first, excessive message passing may lead to over-smoothing in the graph domain, causing node representations from different anatomical regions to become increasingly similar, thereby weakening boundary discrimination; second, deeper graph propagation tends to diffuse affinity information into semantically unrelated or weakly connected regions, which is particularly detrimental when the weak supervision signals provided by CAMs are inherently noisy or incompletely localized. Based on this analysis, we select four layers as the default configuration for the GA-Net, as it consistently delivers strong and statistically superior performance while maintaining stable optimization behavior.
Sensitivity analysis for the supervised segmentation module. To rigorously evaluate the robustness of the exponential moving average (EMA) ensembling mechanism within the supervised segmentation module, we conducted a systematic sensitivity analysis on its two key hyper-parameters: the smoothing coefficient δ and the ensembling interval γ, in strict accordance with the statistical evaluation protocol summarized in Table 3. We tested δ across the values and γ across
using patient-level five-fold cross-validation. The segmentation performance, measured by the DSC on the held-out validation folds, is summarized in Fig 3(b) and 3(c). Statistical analysis using paired Wilcoxon signed-rank tests on patient-level DSC scores confirms the following robustness patterns:
- Smoothing coefficient δ: A broad plateau of high and statistically comparable performance was observed for
. The performance achieved at the optimal setting (
) was not significantly different from that at
or
(both p > 0.05). In contrast, extreme values (
or
) resulted in a statistically significant degradation in performance (p < 0.05). Note: these extreme cases are not shown in the figure.
- Ensembling interval γ: The segmentation performance remained consistently high and without statistically significant differences for
and
(p > 0.05). Significantly lower performance was observed only when the EMA updates were either too frequent (
) or too infrequent (
), with statistically significant differences (p < 0.05). Note: these extreme (> 10) cases are not shown in the figure.
The selected configuration (,
) lies well within these robust, high-performance intervals. This analysis, supported by formal statistical testing, demonstrates that the performance gains introduced by the EMA module stem from its core design principle–iterative prediction stabilization–rather than from meticulous fine-tuning of its hyper-parameters. Overall, the supervised segmentation module exhibits strong inherent stability across a reasonable range of parameter settings.
Sensitivity analysis of loss weight parameters. To rigorously assess the robustness of the proposed multi-task learning framework, we performed a systematic sensitivity analysis on the loss weighting coefficients to
in Eq. (22), in accordance with the statistical evaluation protocol summarized in Table 3. A grid search was conducted within the bounded space
using patient-level cross-validation on the training set only, thereby ensuring that no test data were involved in hyper-parameter selection.
The analysis was designed to examine whether the segmentation performance was critically dependent on a specific combination of loss weights. To this end, we compared the DSC achieved by the final selected configuration with multiple neighboring configurations, in which each individual weight was independently perturbed by
while keeping the remaining coefficients fixed. Statistical analysis based on a series of patient-level paired Wilcoxon signed-rank tests revealed no statistically significant performance differences between the chosen configuration and its immediate variants (all p > 0.05). These results indicate the presence of a broad plateau of stable performance rather than a single, sharply defined optimum.
Ablation studies
To evaluate the contribution of different loss components in the method, we performed ablation experiments with the NIH “TB Portals” [34], ImageCLEF [33], and DeepPulmoTB dataset [7]. The ablation experiment design is shown in Table 4. The baseline method, M-5, incorporates all the loss components and serves as the reference for comparison. The results are summarized in Table 5, which presents the performance of various modules on segmentation and classification tasks. Fig 4 shows a performance comparison between conventional affinity methods and graph-based affinity methods. The corresponding decision-making process and recognition performance of CAM for the ablation experiments M-1 to M-5 are visualized in Fig 5.
A: conventional affinity method. B: graph-based affinity methods.
M-1 (Excluding ): In this experiment, we removed the structural consistency loss
to analyze its impact on model performance.Patient-level paired Wilcoxon signed-rank tests confirmed a statistically significant performance decline in both segmentation (DSC: p < 0.01) and classification accuracy compared to the full model M-5. This indicates that the structural consistency loss is essential for maintaining prediction coherence, particularly in regions with ambiguous boundaries. The CAM visualization in Fig 5 highlights the model’s inability to maintain consistent boundaries, with regions appearing fragmented and misclassified, reinforcing that
significantly enhances structural integrity across anatomical regions.
M-2 (Excluding ): The exclusion of the affinity loss
in M-2 further demonstrates the importance of capturing semantic affinities between regions.Statistical analysis showed a significant decrease in segmentation performance (DSC: p < 0.01), with boundaries becoming less precise.The CAM in Fig 5 illustrates that without
, the model fails to effectively capture relationships between anatomically connected regions, confirming its crucial role in segmenting connected structures.
M-3 (Excluding ): Removing the smoothness loss
in M-3 resulted in a less stable model, especially under noisy or inconsistent supervision. Although the average segmentation accuracy remained relatively high, the performance variation across patients increased significantly, and the model outputs became more erratic (p < 0.05 for increased variance). The CAM in Fig 5 visualizations exhibited more jagged boundaries, indicating that smoothness regularization plays an important role in suppressing noise and stabilizing predictions.
M-4 (Excluding ): The M-4 experiment, which excluded the partial cross-entropy loss (
), showed a statistically significant but modest performance decrease compared to M-5 (DSC: p < 0.05). These results suggest that while the scribble-based module provides beneficial fine-grained boundary information, its absence particularly affects small and irregular lesions. As shown in Fig 5, the CAM for M-4 exhibits less refined segmentation of LCs with weaker boundary definition, emphasizing the contribution of
to capturing fine structural details.
M-5 (Full Model with All Loss Components): The full model (M-5) incorporates all proposed loss components and achieves the best performance across all evaluation metrics.Statistical comparisons against all ablated variants confirmed its significant superiority (all p < 0.01), demonstrating the combined effectiveness of the proposed loss terms. In addition, the CAM visualization in Fig 5 shows well-defined boundaries and accurate anatomical localization. Fig 4 demonstrates that the proposed graph-based affinity algorithm significantly outperforms conventional affinity propagation methods.
Overall, the ablation experiments, supported by rigorous patient-level statistical testing, confirm that each loss component contributes positively to the proposed weakly supervised framework. Specifically, the structural consistency, affinity, and smoothness losses play statistically verifiable roles in preserving boundary coherence, semantic connectivity, and prediction stability, while the scribble-based segmentation module significantly enhances fine-detail recognition for small or irregular lesions.
Moreover, Fig 6 presents a comparative analysis of the training process across five ablation methods (M-1 to M-5). Subfigures (a)–(e) correspond to the loss curves of each method, where the x-axis represents training epochs and the y-axis denotes the loss value. The experiments demonstrate that the training and validation losses for all methods show a convergent trend as epochs increase, and the gap between the validation and training curves remains below 0.15, indicating no overfitting in the models. The magnified views (insets) reveal that the loss values of M-5 in the final stabilization phase (last 50 epochs)—with training loss and validation loss
(mean ± standard deviation)—are numerically lower than those of other methods. Quantitative analysis (Fig 6) shows that M-5 attains a DSC score of 0.457, outperforming methods M-4 to M-1 by 0.06–0.28.
Fig 7 presents a comparative analysis of classification performance across five ablation methods (M-1 to M-5) using confusion matrices, where diagonal elements represent correctly classified samples and off-diagonal elements indicate misclassifications. The experimental results demonstrate that: 1) LCs categories with similar morphological characteristics (e.g., “no cavity” vs. “small cavities <25 mm”, “no cavity” vs. “1–3 cavities”) exhibit significant cross-confusion; 2) The baseline method M-5 achieves optimal classification performance. Fig 6(f) visualizes the ROC curves of each ablation method (M-1 to M-5) in the LCs attribute classification task. The figure shows that the baseline method incorporating all loss functions (M-5) achieved the best AUC in the LCs classification task, with a value of 0.750. These results validate the superiority of our proposed method in TB cavity attribute classification.
Comparison to state-of-the-art
To comprehensively evaluate the performance of our method, we trained and compared several existing medical imaging WSSS methods on the LCs segmentation task, as summarized in Tables 6 and 7. Table 6 presents the benchmark configuration and comparative methods, including their supervision types and backbone architectures. The compared methods include recent state-of-the-art approaches, each designed with varying strategies for weakly supervised segmentation. Notably, some of these methods were originally developed in a 2D manner. To ensure a fair and consistent comparison with our proposed 3D method, we extended these 2D approaches to their 3D counterparts by applying them slice-wise across all dimensions and aggregating the volumetric outputs accordingly. This adaptation enables the evaluation of all methods under a unified 3D experimental setting that aligns with the nature of our input data and task requirements. The quantitative performance comparisons across three public datasets are reported in Table 7.
All methods are re-trained under identical settings with a patch size of and optimized using AdamW.
According to the statistical evaluation protocol defined in Table 3, we analyzed the results using the paired Wilcoxon signed-rank test (significance level p < 0.05) with the patient as the statistical unit. In multiple comparisons with the best baseline method, our proposed method achieved the best performance across all three key evaluation metrics–95HD, IoU, and DSC–and demonstrated statistically significant advantages (p < 0.05) on all three datasets (TB Portals, ImageCLEF, and DeepPulmoTB). Specifically, on the TB Portals dataset, our method achieved results of 27.94 ± 4.63 (95HD), (IoU), and
(DSC), significantly outperforming all existing methods. On the ImageCLEF dataset, the results were
(95HD),
(IoU), and
(DSC), again showing notable improvement over previous methods. On the DeepPulmoTB dataset, our method obtained
(95HD),
(IoU), and
(DSC), maintaining its performance advantage.
In comparison to existing state-of-the-art methods, such as [36], [19], and [15], which show a consistent gap in performance (especially in terms of DSC and IoU), our method stands out in its ability to capture both detailed and large-scale anatomical features. This is reflected in the sharp improvements in the DSC scores across all datasets, with our method consistently achieving the highest scores. Furthermore, Fig 8 illustrates the segmentation results of various methods, where it can be observed that our method achieves higher alignment with the ground truth compared to other approaches. The improved performance can be attributed to the integration of novel components, such as the GA-Net-based affinity learning and region-wise affinity propagation, which allow our model to better capture the relationships between anatomical structures and refine the segmentation boundaries.
C1 - C7 sequentially represent the methods from [36], [19], [37], [15], [24], [38], and [39].
Our method’s superior performance suggests that the proposed method effectively addresses the challenges in weakly supervised 3D medical image segmentation, particularly for LC segmentation. The results indicate that our approach is more capable of handling complex anatomical structures and noisy data, thereby providing a more reliable and accurate segmentation solution compared to the state-of-the-art methods.
Subgroup analysis by lesion characteristics
To better analyze the strengths and limitations of the proposed method, we further conducted qualitative and quantitative subgroup analyses based on lesion size, quantity, morphology, and disease severity. These subgroup definitions are derived from the cavity size and number annotations summarized in Table 1.
Lesion size and quantity. Quantitative analysis shows that the proposed method achieves more pronounced improvements over baseline methods in cases with multiple cavities (more than one cavity) and larger cavity sizes (>25mm), as summarized in Table 8. Within these subgroups, graph-based affinity learning can effectively bridge spatial distances and propagate semantic information between regions associated with relevant anatomical structures, thereby improving the completeness of the segmented regions and reducing fragmentation. In contrast, segmentation becomes more challenging for very small cavities (<25mm) or single-cavity cases, where partial volume effects and low contrast often lead to under-segmentation. Nevertheless, even under such conditions, our method consistently maintains a relative performance advantage over other approaches across all lesion-size-based subgroups.
Average DSC (%) across different lesion subgroups.
Lesion location. Qualitative inspection indicates that cavities located in central lung regions or near major bronchovascular structures are segmented more reliably than those adjacent to pleural boundaries, as summarized in Table 9. Peripheral cavities often exhibit ambiguous boundaries and limited contextual support, which may reduce affinity propagation effectiveness. These failure cases are primarily associated with weak CAM localization rather than errors in the affinity learning module itself.
Lesion morphology. From a morphological perspective, the proposed method performs particularly well on cavities with irregular shapes or fragmented appearances, as summarized in Table 10. By modeling region-level affinities on a superpixel graph, GA-Net is able to capture long-range structural consistency that pixel-level methods often miss. However, thin-walled cavities or cavities with extremely complex internal structures may still suffer from boundary leakage, especially under weak supervision.
Disease severity. When stratifying cases by disease severity using cavity number as a proxy, we observe that the proposed method yields larger relative gains in moderate-to-severe cases (multiple cavities) compared to mild cases (single or no cavity), consistent with the quantitative trends reported in Table 8. This suggests that the proposed approach is particularly effective in scenarios where richer structural context is available, while extremely mild cases remain challenging due to limited lesion evidence.
Overall, these subgroup analyses demonstrate that the proposed method consistently improves segmentation performance across a wide range of lesion characteristics, while also highlighting specific scenarios–such as very small or peripheral cavities–where further improvements are needed.
Computational complexity analysis
To assess the practical feasibility of GA-Net for clinical deployment, we analyze its computational complexity in terms of both theoretical operations and actual runtime performance.
Theoretical complexity: Let be the number of superpixel nodes and
the number of edges in the constructed graph. Each GA-Net layer performs two main operations: (1) edge-wise attention scoring, and (2) node feature aggregation. The attention scoring step computes a scalar for each edge via a learnable projection followed by a dot product, requiring
⋅
operations where d is the feature dimension. The aggregation step then updates each node by a weighted sum of its neighbors, incurring
operations as well. Thus, for L layers, the overall complexity is
. In practice, we keep the graph sparse: adjacency edges are limited to local connectivity (typically 4–8 neighbors per node in 2D, 6–26 in 3D), and long-range kNN edges add a small constant factor (we use k = 10). Hence,
, and the complexity scales linearly with the number of nodes, i.e.,
.
Empirical runtime and memory: We measured the forward-pass time and GPU memory consumption on a representative 3D CT volume (size ) using an NVIDIA RTX 3090Ti. After SLIC superpixel segmentation (which is a pre-processing step and can be efficiently parallelized), the graph contains
nodes and
edges. With a 4-layer GA-Net (d = 256), the affinity inference takes ≈ 42 ms per volume, which translates to ≈ 0.66 ms per slice (96 slices). The peak GPU memory footprint is ≈ 1.2 GB, dominated by the node feature matrices and the attention weights. For comparison, a standard 3D U-Net baseline with similar feature dimensions requires ≈ 120 ms per volume and ≈ 3.5 GB of memory. Thus, GA-Net is ≈ 2.9× faster and uses ≈ 2.9× less memory than the U-Net baseline, making it highly suitable for real-time or near-real-time clinical scenarios.
Comparison with alternative affinity methods: We also compare against two graph-based affinity learning alternatives: (1) a non-parametric method that computes affinities via fixed cosine similarity on encoder features, and (2) a Graph Convolutional Network (GCN) with fixed edge weights. The non-parametric method is slightly faster (≈ 35 ms) but yields significantly lower segmentation accuracy (mIoU drops by 8.2% on our validation set). The GCN baseline has similar runtime to GA-Net (≈ 45 ms) but lacks the adaptive attention mechanism, leading to a 4.5 % mIoU reduction. Hence, GA-Net achieves a favorable trade-off between efficiency and accuracy.
Clinical deployment considerations: The linear scaling with respect to the number of nodes ensures that the model remains efficient even for high-resolution volumes. In a typical clinical workflow, the entire pipeline–including superpixel generation, GA-Net inference, and random-walk refinement–runs in under 200 ms per volume on a single GPU, meeting the real-time requirements for interactive segmentation tools. Furthermore, the model can be easily optimized via TensorRT or ONNX runtime for further speedup on embedded devices.
In summary, GA-Net introduces minimal computational overhead compared to conventional segmentation networks, while providing the structured reasoning necessary for robust affinity learning. Its low latency and memory footprint make it a practical choice for clinical deployment, where both accuracy and efficiency are critical.
Discussion
The method proposed in this study represents a meaningful advance in the task of weakly supervised semantic segmentation of lung cavities in 3D medical images. Experimental results demonstrate that, compared to existing baseline methods, our model achieves significant relative performance improvements across multiple datasets (e.g., an average relative improvement of approximately 5% in DSC). This improvement validates the effectiveness of the proposed graph affinity learning and scribble-guided optimization mechanisms in leveraging weak supervision signals.
We note that, while the relative improvements are significant, there remains room for enhancement in the absolute performance levels of segmentation for this task (e.g., an average DSC of approximately 45.4% and IoU of approximately 27.3%). This reflects the inherent challenges of accurately segmenting morphologically complex and boundary-ambiguous lung cavities under the weakly supervised setting using only image-level labels and sparse scribbles. The current performance bottlenecks may stem from several aspects: firstly, the quality of the initial pseudo-labels is highly dependent on the CAM’s ability to localize discriminative regions, which may be incomplete when the target exhibits high internal heterogeneity; secondly, the weak supervision signals themselves lack exhaustive boundary information, imposing a theoretical upper limit on the model’s ability to learn fine contours; finally, the substantial variations in lesion size, morphology, and contrast across different datasets also pose difficulties for model generalization.
Regarding the prospects for clinical translation, we consider the current work primarily as a methodological validation. Although the results are encouraging, a series of more in-depth studies and validations are required before it could be directly deployed in clinical decision support systems. Examples include external validation on larger-scale, multi-center, prospectively collected datasets; further quantitative analysis of the model’s performance across different subgroups (e.g., based on cavity size, location, or etiology); and conducting necessary clinical trials to evaluate its practical impact on clinical workflows and decision-making. Furthermore, the model’s computational efficiency is also a consideration for practical deployment, and future work could explore model lightweighting and inference acceleration techniques.
Overall, this study indicates that graph-based affinity learning holds considerable potential for enhancing segmentation performance under weak supervision and offers a promising research direction for reducing annotation costs in 3D medical image analysis. Future work will focus on improving absolute segmentation accuracy, enhancing cross-dataset generalization capability, and further evaluating its clinical relevance through larger-scale validation.
Conclusion
In this paper, we have proposed a novel weakly supervised method for 3D semantic segmentation of lung cavities, achieving state-of-the-art performance across multiple benchmark datasets. Our method, which integrates graph-based affinity learning, region-wise affinity propagation, and scribble-based semantic segmentation, effectively addresses the challenges of weak supervision in medical image segmentation. The comprehensive ablation study demonstrated the critical role of each component in improving the segmentation accuracy, particularly in complex anatomical structures.While there are some limitations, such as high computational costs, our method shows potential as a reliable tool for the automated analysis of 3D medical images. Future improvements in annotation strategies and computational efficiency will further enhance its applicability and robustness in clinical settings.
References
- 1. Dartois VA, Rubin EJ. Anti-tuberculosis treatment strategies and drug development: challenges and priorities. Nat Rev Microbiol. 2022;20(11):685–701. pmid:35478222
- 2. Tan Z, Madzin H, Norafida B, Rahmat RWO, Khalid F, Sulaiman PS. SwinUNeLCsT: global–local spatial representation learning with hybrid CNN–transformer for efficient tuberculosis lung cavity weakly supervised semantic segmentation. Journal of King Saud University - Computer and Information Sciences. 2024;36(4):102012.
- 3.
Tang Y, Yang D, Li W, Roth HR, Landman B, Xu D, et al. Self-supervised pre-training of swin transformers for 3d medical image analysis. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022. p. 20730–40.
- 4.
Zhou HY, Guo J, Zhang Y, Yu L, Wang L, Yu Y. nnformer: Interleaved transformer for volumetric segmentation. arXiv preprint 2021. https://doi.org/arXiv:2109.03201
- 5.
Hatamizadeh A, Tang Y, Nath V, Yang D, Myronenko A, Landman B, et al. Unetr: Transformers for 3d medical image segmentation. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision; 2022. p. 574–84.
- 6.
Wang W, Chen C, Ding M, Yu H, Zha S, Li J. Transbts: multimodal brain tumor segmentation using transformer. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021 : 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part I 24. Springer; 2021. p. 109–19.
- 7. Tan Z, Madzin H, Norafida B, ChongShuang Y, Sun W, Nie T, et al. DeepPulmoTB: a benchmark dataset for multi-task learning of tuberculosis lesions in lung computerized tomography (CT). Heliyon. 2024;10(4):e25490. pmid:38370224
- 8.
Wu T, Huang J, Gao G, Wei X, Wei X, Luo X, et al. Embedded discriminative attention mechanism for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2021. p. 16765–74.
- 9. Zhang D, Li H, Zeng W, Fang C, Cheng L, Cheng M-M, et al. Weakly supervised semantic segmentation via alternate self-dual teaching. IEEE Trans Image Process. 2025;34:3086–95. pmid:38117621
- 10. Li R, Mai Z, Zhang Z, Jang J, Sanner S. TransCAM: transformer attention-based CAM refinement for Weakly supervised semantic segmentation. Journal of Visual Communication and Image Representation. 2023;92:103800.
- 11.
Lee H, Jeong WK. Scribble2label: Scribble-supervised cell segmentation via self-generating pseudo-labels with consistency. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2020 : 23rd International Conference, Lima, Peru, October 4–8, 2020, Proceedings, Part I. 2020. p. 14–23.
- 12. Liu X, Yuan Q, Gao Y, He K, Wang S, Tang X, et al. Weakly supervised segmentation of COVID19 infection with scribble annotation on CT images. Pattern Recognit. 2022;122:108341. pmid:34565913
- 13.
Lee J, Yi J, Shin C, Yoon S. Bbam: Bounding box attribution map for weakly supervised semantic and instance segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2021. p. 2643–52.
- 14. Yi S, Ma H, Wang X, Hu T, Li X, Wang Y. Weakly-supervised semantic segmentation with superpixel guided local and global consistency. Pattern Recognition. 2022;124:108504.
- 15.
Kweon H, Yoon KJ. From sam to cams: exploring segment anything model for weakly supervised semantic segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024. p. 19499–509.
- 16.
Chen Q, Hong Y. Scribble2d5: Weakly-supervised volumetric image segmentation via scribble annotations. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer Nature Switzerland; 2022. p. 234–43.
- 17. Li Z, Xia Y. Deep reinforcement learning for weakly-supervised lymph node segmentation in CT images. IEEE J Biomed Health Inform. 2021;25(3):774–83. pmid:32749988
- 18. Sun W, Feng X, Liu J, Ma H. Weakly supervised segmentation of COVID-19 infection with local lesion coherence on CT images. Biomed Signal Process Control. 2023;79:104099. pmid:35996574
- 19. Lu F, Zhang Z, Liu T, Tang C, Bai H, Zhai G, et al. A weakly supervised inpainting-based learning method for lung CT image segmentation. Pattern Recognition. 2023;144:109861.
- 20.
Ahn J, Cho S, Kwak S. Weakly supervised learning of instance segmentation with inter-pixel relations. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2019. p. 2209–18.
- 21. Wang X, Liu S, Ma H, Yang M-H. Weakly-supervised semantic segmentation by iterative affinity learning. Int J Comput Vis. 2020;128(6):1736–49.
- 22. Früh M, Fischer M, Schilling A, Gatidis S, Hepp T. Weakly supervised segmentation of tumor lesions in PET-CT hybrid imaging. J Med Imaging (Bellingham). 2021;8(5):054003. pmid:34660843
- 23. Li X, Li B, Yin H, Xu B. An automatic random walker algorithm for segmentation of ground glass opacity pulmonary nodules. J Healthc Eng. 2022;2022:6727957. pmid:36212245
- 24. Du H, Dong Q, Xu Y, Liao J. Weakly-supervised 3D medical image segmentation using geometric prior and contrastive similarity. IEEE Trans Med Imaging. 2023;42(10):2936–47. pmid:37093730
- 25.
Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for discriminative localization. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016. p. 2921–9. https://doi.org/10.1109/cvpr.2016.319
- 26.
Ru L, Zhan Y, Yu B, Du B. Learning affinity from attention: End-to-end weakly-supervised semantic segmentation with transformers. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022. p. 16846–55.
- 27. Huang Z, Wang M, Hou M, Yu Z, Wang S, Li X, et al. BACFormer: a robust boundary-aware transformer for medical image segmentation. Knowledge-Based Systems. 2026;335:115209.
- 28. Han W, Kang S, Choo K, Hwang SJ. Complementary branch fusing class and semantic knowledge for robust weakly supervised semantic segmentation. Pattern Recognition. 2025;157:110922.
- 29. Achanta R, Shaji A, Smith K, Lucchi A, Fua P, Süsstrunk S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans Pattern Anal Mach Intell. 2012;34(11):2274–82. pmid:22641706
- 30. Wang Y, Yang R, Sun L. A novel structure adaptive discrete grey Bernoulli model and its application in renewable energy power generation prediction. Expert Systems with Applications. 2024;255:124481.
- 31. Chang W, D’Ascenzo N, Antonecchia E, Li B, Yang J, Mu D, et al. Deep denoiser prior driven relaxed iterated Tikhonov method for low-count PET image restoration. Phys Med Biol. 2024;69(16):10.1088/1361-6560/ad67a3. pmid:39053501
- 32. Müller R, Kornblith S, Hinton GE. When does label smoothing help?. Advances in Neural Information Processing Systems. 2019;32.
- 33.
Kozlovski S, Liauchuk V, Dicente Cid Y, Kovalev V, Müller H. Overview of ImageCLEFtuberculosis 2021 : CT-based tuberculosis type classification. In: Proceedings of the CLEF 2021 Conference and Labs of the Evaluation Forum. 2021.
- 34. D’Souza NS, Wang H, Giovannini A, Foncubierta-Rodriguez A, Beck KL, Boyko O, et al. Fusing modalities by multiplexed graph neural networks for outcome prediction from medical data and beyond. Med Image Anal. 2024;93:103064. pmid:38219500
- 35. Yushkevich PA, Yang Gao, Gerig G. ITK-SNAP: an interactive tool for semi-automatic segmentation of multi-modality biomedical images. Annu Int Conf IEEE Eng Med Biol Soc. 2016;2016:3342–5. pmid:28269019
- 36.
Laradji I, Rodriguez P, Manas O, Lensink K, Law M, Kurzman L. A weakly supervised consistency-based learning method for covid-19 segmentation in ct images. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 2021. p. 2453–62.
- 37. Lin L, Liu Y, Wu J, Cheng P, Cai Z, Wong KKY, et al. FedLPPA: learning personalized prompt and aggregation for federated weakly-supervised medical image segmentation. IEEE Trans Med Imaging. 2025;44(3):1127–39. pmid:39423080
- 38. Lei W, Su Q, Jiang T, Gu R, Wang N, Liu X, et al. One-shot weakly-supervised segmentation in 3D medical images. IEEE Trans Med Imaging. 2024;43(1):175–89. pmid:37440388
- 39. Wang H, Huai L, Li W, Qi L, Jiang X, Shi Y. WeakMedSAM: weakly-supervised medical image segmentation via SAM with sub-class exploration and prompt affinity mining. IEEE Trans Med Imaging. 2025;44(7):2795–807. pmid:40063440