Ultrasound segmentation is a challenging problem due to the inherent speckle and some artifacts like shadows, attenuation and signal dropout. Existing methods need to include strong priors like shape priors or analytical intensity models to succeed in the segmentation. However, such priors tend to limit these methods to a specific target or imaging settings, and they are not always applicable to pathological cases. This work introduces a semi-supervised segmentation framework for ultrasound imaging that alleviates the limitation of fully automatic segmentation, that is, it is applicable to any kind of target and imaging settings. Our methodology uses a graph of image patches to represent the ultrasound image and user-assisted initialization with labels, which acts as soft priors. The segmentation problem is formulated as a continuous minimum cut problem and solved with an efficient optimization algorithm. We validate our segmentation framework on clinical ultrasound imaging (prostate, fetus, and tumors of the liver and eye). We obtain high similarity agreement with the ground truth provided by medical expert delineations in all applications (94% DICE values in average) and the proposed algorithm performs favorably with the literature.
Citation: Ciurte A, Bresson X, Cuisenaire O, Houhou N, Nedevschi S, Thiran J-P, et al. (2014) Semi-Supervised Segmentation of Ultrasound Images Based on Patch Representation and Continuous Min Cut. PLoS ONE 9(7): e100972. https://doi.org/10.1371/journal.pone.0100972
Editor: Arrate Muñoz-Barrutia, University of Navarra, Spain
Received: February 17, 2014; Accepted: June 1, 2014; Published: July 10, 2014
Copyright: © 2014 Ciurte et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: This work is supported by PRODOC Project of Technical University of Cluj-Napoca and by the Center for Biomedical Imaging (CIBM) of the Geneva-Lausanne Universities and EPFL, and the foundations Leenaards and Louis-Jeantet, and by the FNS-205321-141283 and CTI-13741.1 funds. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Ultrasound (US) imaging provides a high image resolution and simplicity of use at much lower cost in comparison with other medical imaging modalities such as Magnetic Resonance imaging (MRI) or Computer Tomography (CT). These advantages make ultrasound a main image modality in medicine and its use becomes of particular interest in areas such echocardiography –, obstetrics and gynecology , breast cancer  and intravascular diseases . Indeed, US imaging is more and more used for image-guided therapy planning and navigation, and computer aided diagnosis. To this end, the development of efficient US image segmentation techniques of biological structures (e.g. different organs, heart chamber, fetus) or focal diseases (e.g. tumors, cysts) is needed.
US segmentation is very challenging due to the inherent speckle and some artifacts like shadows, attenuation and signal dropout. This often leads to weak (or missing) edges and also to the presence of fake edges, making standard unsupervised segmentation methods fail. Indeed, segmentation methods succeed on US imaging only when making use of application-specific constrains or priors. Several types of priors have been suggested in the literature : those coming from time or functional data and those based on shape or imaging physics. In this work, we segment 2D B-mode US images without contrast enhancement. In consequence, we will not discuss the use of time-based constraints from different imaging frames or functional-based priors provided by contrast agent. However, such kind of priors could be included in our formulation if needed.
Anatomical shape priors are powerful for US image segmentation dealing very well with shadows and weak edges and they have been successfully used for different anatomical structures segmentation such as the heart , , the prostate , , the breast  or the kidney . These priors can be encoded in the form of statistical shape models, usually derived from large data set segmentations, and they are often computationally expensive. Unfortunately, they are of limited use for pathological cases due to high variation in structure, shape, size and localization of lesions. Shape priors can be encoded in a simpler manner as for instance imposing a smooth boundary . We will have such regularization in our approach.
Priors can also be related to the imaging physics, i.e. related to the observed intensities of the ultrasound image. For example, we consider intensity-based priors or learned texture as specific parametric intensity models to characterize the observed data. In this context, two different philosophical branches can be identified: those that consider speckle as information and those that consider it as noise. Since the gray level intensities in US images reflect the tissue density, some approaches use denoising filters in order to reduce the speckle structure and smooth the images (e.g. ). On the contrary, other approaches benefit of the texture information contained by speckle. As proved by Oosterveld et al.  by realistic simulations, the statistical and speckle characteristics of echographic texture change according to the density and spatial distribution of scatterers within the resolution cell.
A wide range of image features are used for modeling US observed data (see  for an extended review). Intensity features are often used by means of analytical models of the gray-level distribution (Rayleigh distribution being the most used model  but also Exponential, Gamma and Gaussian). Note that all these models are well-suited for the received signal while, in practice, clinical ultrasound devices log-compress the signal before visualization . The main drawbacks of these gray-level distribution models are related to distribution parameter estimation and to its dependency on the imaging system settings such as dynamic range and gain. Intensity gradient features can also be used . However, they strongly assume that there are homogeneous regions and thus, a very low level of speckle noise. Texture features have also proved to be successful in US segmentation, particularly statistical patterns of the intensity due to their advantage of being independent of the imaging system physics .
Contribution of this work
We consider in this work an alternative approach for characterizing the ultrasound data. Rather than using analytical distribution models or texture patterns to represent ultrasound images, we make use of graph of intensity patches as image representation. In contrast with most intensity or texture-based methods, the use of patches as image features allows us to be more independent of the imaging system and no assumption on the echogenicity of the object has to be done. Our choice of patch feature is also supported by previous works in the literature. Patch features were first introduced for texture synthesis  and image denoising  for natural images. In , Coupé et al. extended the non-local means filter  to reduce speckle noise in US images, by defining a particular similarity measure between patches. In , , the patch-based approach was also successfully used for spatio-temporal registration and for motion or/and elasticity estimation in US image sequences of the heart. As far as we know, our preliminary work , published in 2011, was the first paper that introduces the use of intensity patches in US segmentation. Besides, we use the Pearson distance between patches as it proves to be a robust distance to speckle, as shown by .
Given the graph US image representation, we then address the segmentation problem with an efficient and interactive extraction algorithm of the foreground object, where the background cannot be trivially subtracted. Our segmentation method is thus fundamentally semi-supervised, that is, initial labels are defined on the image, acting as soft priors. Interactive soft priors were introduced in Computer Vision with user-assisted segmentation algorithms such as Interactive Graph Cuts , Lazy Snapping  and GrabCut ) and also in Medical Imaging (like CT  and US  segmentation. This quick and easy way to interact with images can provide the priors needed to make our segmentation accurate, robust and applicable to different kind of targets or imaging parameters. From the interaction point of view, our algorithm may be equivalent to a large variety of US segmentation methods proposed in the literature. Several state-of-the-art methods  require initial clicks or other types of interaction, like defining a region of interest or much tedious manual training. Moreover, depending on the application's specificities, the label initialization can be automatized, as in  for retinoblastoma segmentation.
Our segmentation method is based on the patch-based continuous graph cut approach for natural image segmentation introduced in . Continuous graph cut methods have seen a rapid development over the recent years –. These methods find their theoretical roots in Strang  but the interest of applying them to real-world applications like medical imaging and computer vision has been triggered only recently. Nevertheless, continuous graph cut methods are quite attractive and offer new features that we will discuss in a further section.
A preliminary version of this work was presented in . The main differences and improvements of the proposed work with respect to our previous work are as follows. Firstly, we introduce a new minimization algorithm that speeds up by at least an order of magnitude our previous optimization algorithm proposed in . Secondly, we perform several tests of the proposed ultrasound segmentation methodology to identify clinical targets, namely, prostate, fetus, liver tumors and eye. Eventually, we carry out a thorough study of the robustness of our segmentation algorithm with respect to initialization.
As shown in , most US segmentation methods are usually limited to a specific target (as for instance the endocardial border , , , , breast mass , –, prostate , , , , and liver , ). The major advantage of our segmentation framework, as regards the state-of-the-art, is its flexibility and easiness to use, while obtaining equivalent or better accuracy. However, this prevents us from selecting one approach in the literature for comparison that would not be equivalent in terms of flexibility. We will thus focus on one clinical context for method comparison, the segmentation of the prostate. This is one of the main areas of application of US in cancer treatment . We have chosen the most recent, and methodologically close, semi-supervised approach presented in the literature , that makes use also of graph theory and soft priors using initial labels.
We now summarize the main methodological contributions of the proposed US segmentation method:
- An efficient representation of US images based on graph of intensity patches (naturally adapted to any echogenicity and imaging systems),
- A fast minimization algorithm for the minimum cut problem,
- An easy use of soft priors based on interactive user labels (unlike hard priors such as shape or temporal constrains),
- A study of accuracy and reproducibility of our algorithm with respect to different initial labels and users;
and the main application-oriented contributions are:
- A flexible framework, applicable to different US segmentation problems (eye, liver, prostate and fetus),
- A high accuracy of segmentation results as compared to manual delineations of expert raters, with mean Dice values around in all data sets.
A related work on ultrasound image segmentation were recently introduced in ,  in the context of echocardiographic sequences. This work also promotes the use of intensity patches for ultrasound segmentation. More precisely, the authors of ,  define a sophisticated image representation model with multiscale signal analysis and sparse coding technology, which make the image representation rich but also time-consuming. We directly use image patches as image representation and show that this simple image representation is quite flexible and accurate to deal with multiple segmentation scenarios. Besides, ,  make use of the level set method (LSM) as segmentation method. The LSM is a PDE-based segmentation technique, which speed is limited by the Courant-Friedrichs-Lewy (CFL)'s condition. Discrete and continuous graph cut techniques (like the one introduced in this work) are more recent segmentation techniques and are not limited by the CFL condition, thus faster than the LSM (at least by an order of magnitude).
The proposed paper is organized as follows. In the first section, our mathematical framework and minimization scheme are presented. In the second section, the parameter setting and datasets are introduced. In the third section, we study the accuracy, repeatability and computational speed with respect to the initialization step. Then, quantitative results are presented for liver, eye tumors, fetal head estimation, and prostate. Eventually, a discussion is presented in the last section.
The general block diagram of the proposed US segmentation method is shown in Figure 1. With a given US image as an input, the method first models the image with a graph representation of patches; next, the user provides markers inside (and outside, if needed) the object as initialization, serving as method initialization; finally, the novel numerical scheme for the minimization process is applied. Note that, in practice, if the segmentation result is not satisfying, our framework offer the possibility to re-initialize the labels and segment again, without re-computing the image modeling step. We will denote our new fast Patch-based Continuous Min-Cut segmentation by fP-CMC to distinguish it from our previous version P-CMC in .
Modeling of US images
Let an image be defined as a vector in where is the number of pixels in the image. In this work, we represent US images as a graph of image patches. Therefore we introduce a graph , where the set contains the nodes of the graph, the set contains the edges (i.e. the connections between the nodes) and is the weight matrix that encodes the similarity between two points ( is small when is different from and is large when is similar to ). Each node of the graph represents a pixel in the image The following graph weight is considered:(1)where expresses the proximity between the image descriptors located at and , is a neighborhood window centered at of size , and the parameter is the scaling parameter of the weight matrix (using a standard Gaussian kernel with 0-mean and variance ). In our case, the image descriptors and are image patches centered at and , respectively. Note that the proposed segmentation framework is flexible enough to support any types of image descriptor. A standard distance that measures the similarity between image patches is the norm, a.k.a. the mean square difference, as used in  for natural image segmentation. Our patch feature is more adapted for US image segmentation than most texture descriptors used in the literature in several ways: it does not need multiple resolution characterization, it does not need any adjustment for ultrasound imaging (except from the patch size that can be easily set), it has no trouble in segmenting small areas (as some texture-based methods do) and, it is extremely easy to be computed.
Speckle model and Pearson distance
In the context of US imaging, patch distance is hard and critical to define. We need to consider the complex image formation of the US images such as local correlation due to periodic arrangements of scatterers and finite beam width, envelop detection and logarithm amplification of radio-frequency signals, and additive Gaussian noise of sensors. Consequently, we choose the following speckle model, that was proved in  to fit well the log-compressed US images:(2)where is the original image, is the observed image and is a Gaussian noise. Thus, for each pixel can be made the assumption that(3)
An extensive study was performed by Coupé et al.  that proved the higher performance of the Pearson distance against the L2 norm to measure the similarities between patches in US images. Thus, we choose to use the Pearson distance between patches for segmentation since it is better adapted to the speckle:(4)where , where is the number of pixels in the patch centered at and is the intensity of element of the patch.
Continuous graph cut model.
An efficient approach to solve the US segmentation problem is to cast the problem as a continuous graph partitioning problem such as , . In this work, we design a continuous graph cut model to carry out the US segmentation task with soft label priors.
Continuous graph cut methods, such as see –, have seen a rapid development in the recent years based on new attractive properties. Different from discrete graph cut methods such as , , , , which are generaly based on combinatorial optimization techniques, continuous graph cut techniques are fundamentally different: they are built on continuous tools like elliptic Partial Differential Equation (PDEs), variational methods and continuous convex optimization techniques like augmented Lagrangian, Uzawa-type primal-dual, iterative shrinkage techniques, etc. Mathematical fields like Functional Analysis offer strong tools to analyze the well-posedness of these continuous graph cut techniques. The main advantages of the continuous formulation of the graph cut methods are 1) sub-pixel accuracy, 2) easy to code (a few lines of Matlab) and 3) easy to parallelize with significant speedups.
Let us now introduce our continuous graph cut methodology, which benefits from these mentioned new features. Graph partition methods aim at segmenting a graph into two subsets, and , such that and the inter-similarity between and being minimized. Partitioning a graph into two sets and can be achieved by minimizing the cut operator defined as:(5)which is equivalent to:(6)when is an indicator function of the set A. More specifically, we have with is the element of the vector and is the number of pixels in the image. Function is an indicator function of the set when for and for It is known that minimizing the cut operator (5) is equivalent to minimizing the graph-based norm as long as binary indicators, i.e. , are considered . The graph-based norm is defined as:(7)
Observe that the weighted graph Laplacian naturally corresponds to a finite difference approximation of the continuous Laplacian operator and it also corresponds to the non-local operator of diffusion on graph.
Segmentation with labels.
The proposed segmentation algorithm aims at minimizing the cut operator (or graph-based norm) given some labels to identify the objects of interest and the background. The object labels are defined as for and the background labels are defined as for . We now formulate the segmentation method as a discrete minimization problem:(8)where is defined in (7) and are provided by the user. As Shi and Malik observed in , minimizing the cut can favor small sets. An easy way to overcome this issue, while smoothing the irregularities along the cut boundary, is to use the total variation (TV) norm (i.e. the norm of the gradient) as follows:(9)with , where are the discrete spatial derivative operators in the x,y-directions. The discrete minimization problem (9) is a combinatorial problem difficult to solve because the set of minimization is the binary function () which is not convex. The natural approach is to relax the binary constraint to the closest convex set which is naturally . We thus consider the following continuous minimization problem:(10)which is convex, therefore providing a unique solution for any initial condition. Another nice consequence of (10) is the opportunity to develop efficient continuous minimization schemes, based on recent development for optimization (for the compressed sensing field). The proposed segmentation model (9) was already introduced in  (based on ). However, we will present in the next section a novel and (much) faster minimization algorithm to solve (10).
Efficient minimization algorithm
This section introduces a fast minimization algorithm for (10) based on augmented Lagrangian method and splitting technique such as –. The minimization problem (10) is equivalent to (splitting step):(11)
Adding new variables by a splitting step is a standard approach that can solve a difficult minimization problem by equivalently solving easier sub-minimization problems. Indeed, the constrained minimization problem (11) can be solved by the following iterative unconstrained minimization problems (augmented Lagrangian step):(12)
The Euler-Lagrange equation of (13) is:(14)where is the graph Laplacian. The linear system of equations (14) can be solved efficiently with a conjugate gradient method (with e.g. Matlab). Then, the constraints and , are simply imposed to the solution of (14). The second sub-minimization problem to solve is:(15)
The Euler-Lagrange equation of (15) is:(16)which can be solved quickly with fast Fourier transform (FFT) or discrete cosine transform (DCT) depending on the boundary condition. The third and last sub-minimization problem is:(17)which solution is given by soft-thresholding :(18)
The pseudo-code is given in Table 1. to summarize the proposed algorithm. Finally, the stopping condition is chosen to be and in all experiments.
A detailed study of the parameters , and has previously been done in the context of image denoising for natural  and ultrasound images . We briefly re-visit here after the parameter setting in the context of B-mode US segmentation.
Searching window size () - It has a great impact on the computational time. The size of searching window is also related to the “non-local character” of the method, since the similar patches are chosen inside this window and not only from the nearest neighbors.
Patch size () - The patch size is known to be related to the texture pattern size and to the target object scale , . In ultrasound, texture pattern is provided by speckle. High values of imply an increased computational time of weighted matrix and the segmentation becomes coarser and a loss of details is noticed. Thus, we set to ensure the highest precision and a fast computation .
Scaling parameter () - It stands for the typical distance between similar patches, which depends on the noise level . For the proposed speckle model, according to Eq. (3), we have a level of noise equal to . Thus, the selection of parameter depends on the distribution of the gray levels inside the reference patch (). That implies that also depends on speckle characteristics such as speckle size and if it is fully developed or not. Based on the same assumption defined in Eq. (3), the level of noise has a high variation because of the speckle appearance. This leads to a low power of discrimination between two different tissues with close density properties (meaning a low contrast in US image). A value of performs goods results in case of US images, with a higher sensitivity for the mentioned cases of tissues with close density properties.
Regularization weight () - It controls the smoothness of the contour and eliminates the misclassification of small subsets of pixels resulted by the min-cut algorithm. Optimal results in ultrasound segmentation are performed for values between and . Larger value, smoother the contour is. The sub-parameters and occurred in the augmented Lagrangian step (Eq. 12) are both fixed to .
For all the tests and clinical applications presented in this paper we used the following setting: , , and .
Our data set contain a wide variety of ultrasound images in order to proof its flexibility and performance for different imaging systems, image resolution, speckle size and targets. It contains a total number of images from which 20 ophthalmic, 30 liver, 22 fetal and 6 prostate images. Table 2 summarize the characteristics of each data set. Besides the US images, we dispose of manual delineation performed by medical experts for the ophthalmic, liver and prostate data sets and of manual biometric measurements for the fetal data set, which were used as ground truth in our tests.
The patient information from all data used in our study was anonymized and de-identied by physicians prior to our analysis. All studies presented are approved by the corresponding committee/institutional review board: Switzerland (Cantonal Research Ethics Committee of Vaud), France (Comite de Protection des Personnes), Romania (local Ethical Committee of the University of Medicine and Pharmacy Cluj-Napoca) and Canada (Health Science Research Ethics Board at Western University).
Ultrasound is still in its early years as regards ophthalmology. However, its use is constantly increasing ,  for diagnosis, planning, therapy and follow up of treatments. We present the segmentation of B-scan US of the retinoblastoma. The clinical value of this work is related to a larger project that aims at improving the radiotherapy planning and treatment of retinoblastoma in childhood by fusion of CT, ultrasound and fundus of photograph . Our clinical eye data set contains ophthalmic US images which are 2D slices of the 3D volumes from eyes with retinoblastoma (including both calcified and non calcified tumors). US eye images were acquired at Jules Gonin Hospital, Lausanne, Switzerland with Ophthalmic Technologies Inc. (OTI), having an isotropic resolution of mm2 and image size of pixels.
Liver ultrasound aims at finding abnormalities, such as scarring (cirrhosis), masses (both cancer and non cancer) and blockage of the blood vessels. These findings help determine the diagnosis and therapy and also whether the patient is good or not as transplant candidate. In clinical practice, liver diagnosis requires additional examination with other invasive methods like biopsies, with the associated morbidity and mortality risk. Therefore, the automated segmentation of ultrasound images would provide a reliable non-invasive and quantitative approach in diagnosing liver diseases.
We will present the lesion segmentation of US liver imaging from different patients diagnosed with hepatocarcinoma (HCC). These application does not aim at illustrating any clinical value but a quantitative validation thanks to the available manual delineations. The US liver images were acquired at Medical Clinic, Cluj-Napoca, Romania using Logiq7 system at a frequency of MHz, with isotropic resolution of mm2 ( cm) and mm2 ( cm) respectively and image size of pixels. The data set contains two images per patient, one for each resolution, summing up a number of images.
Ultrasound imaging is the gold standard modality for exploration, biometric measurements and diagnosis in fetal imaging. In this context, the development of efficient US image segmentation techniques of biological structures such as the head, femur or abdomen of the fetus are crucial towards a high quality perinatal follow up and diagnosis. Here we aim at segmenting the head of the fetus to estimate later on the head circumference (HC) as an indirect measure of fetal growth. Our data set contains 22 US images from 22 different fetuses ranging from 24 to 35 weeks of gestational age (GA). Images were acquired in clinical practice at Hôpital Femme Mère Enfant, Lyon, France with Siemens Medical Systems at with spatial resolution of isotropic pixels between to . Let us note that only four fetuses can be considered as healthy while the rest are suspected of developmental brain delay. Inner head circumference computed by expert radiologist will be used as gold standard. Our previous method P-CMC  was evaluated on a different dataset to estimate the outer HC in .
Transrectal ultrasound (TRUS) is a key tool for prostate cancer diagnosis. Prostate volumes and boundaries are essential biomarkers in the diagnosis, treatment, and follow up of prostate cancer . This has encourage many researches to develop segmentation tools for prostate boundary detection . The testing data set contains 6 prostate US images, the ones introduced and used for tests in .
Let us note that the software and data set for testing presented in this paper will be made publicly available at http://www.unil.ch/mial/page86599.html website upon acceptance of the manuscript.
The overlap measures that we use to estimate the agreement between the segmentation result and the ground truth (GT) are: Dice coefficient metric (DCM)  and area overlap (AO, or accuracy overlap) . They are defined as follows:where by and we refer to the set of segmented, respectively ground truth points, and
By we denote the set cardinal and TP, FN and FP are the true positive, false negative and false positive sets respectively. Added to overlap measures for segmentation agreement, speed in number of iterations and computational time are also shown. Our algorithm is implemented in Matlab and it was tested on an INTEL Core 2 Duo, 2.66 GHz, 2 GB of RAM.
Robustness to initialization
Our segmentation method is designed in an interactive framework which allows the user to easily and quickly initialize the algorithm for any target of interest. Let us remind that, depending on the specific application, the initialization process can be easily automatized . However, we promote here some initial user interaction in favor of a great gain in flexibility: we will be able to segment many different targets. Having some initial interaction is not new in US segmentation. A large range of state-of-the-art methods must define initial clicks and even heavier interactions (please refer to Table II and III in  as regards interaction type for breast and prostate segmentation respectively). In this sense, our algorithm is not different from algorithms in the literature.
To this end, we present a robustness study with different type of initializations: simple straight lines, one click-drag and drop for generating predefined shapes like ellipses or circles, or freely drawing scribes (as they have prove to be largely used in interactive segmentation applications , . The initialization requires the selection of at least one object, and, optionally, of one or more background areas.
We test the robustness of our method for both eye and prostate imaging. Some examples of different types of initial labels (lines, ellipses and free hand) are shown in Figure 2. Moreover, three different users have drawn the initial labels to study also the inter-user variability. Results are shown in the boxplots of Figure 3. For a given user, DCM is in average for both eye and prostate imaging with and of variance respectively. Therefore, this demonstrate a high robustness as regards different types of initializations. Moreover, inter-user variability for a given label is also highly robust (around ). We applied the Wilcoxon two-sided rank sum test, a non-parametric alternative to paired-student test since it does not make any assumptions regarding the distributions of the data population, on DCM statistics. All tests accepted the null hypothesis at 5% significance level, that is, differences between each boxplot are not statistically significant. Therefore we can conclude that all types of initialization are valid for segmentation and that results are repeatable by different users.
Yellow and light blue correspond to foreground and background labels respectively.
central red mark is the median, edges of the box are the 25th and 75th percentiles, whiskers extend to extreme values, and outliers are plotted by a red cross.
Additional study concerning the spatial filling of the initial labels can be found in File S1 attached to the manuscript.
Quantitative assessment is presented for the eye, the liver, the prostate, and the fetus. The color code used in figures 4, 5, 7 and 9 is as follows: foreground and background labels are in yellow and light blue respectively, fP-CMC segmentation is in solid red contour, and ground truth is in transparent green (except for figure 7 where ground truth segmentation is not available, only the fetal head circumference value).
ground truth is in transparent green, fP-CMC segmentation is in red, foreground and background labels are in yellow and light blue respectively.
Image resolution is for (a), (b) and (c), and for (d), (e) and (f). Ground truth is in transparent green, fP-CMC segmentation is in red, foreground and background labels are in yellow and light blue respectively.
Segmentation of eye tumors.
Segmentation results of US image segmentation are shown in Figure 4 (input images in the first line, segmentation results in the second one and zoom on the results in the third one). The Dice coefficient metric and computational time, in comparison with the previous P-CMC method , are presented in Table 3. DCM is with a standard deviation of was obtained in average for the images of the eye data set. The speed-up of the novel method is clear visible from the reduced time of convergence and number of iterations presented. We measure the intra- and inter-observer variability of two raters that manually delineated the ground truth retinoblastoma. The inter-observer Dice average and variance is of and the intra-observer variability is of . The fP-CMC Dice average of is good as compared to the variability measurements.
Segmentation of liver tumors.
A region of interest of was defined from the original images (first row of Figure 5). Segmentation results are shown in second row of Figure 5 for different resolutions: (a) to (c) have a resolution of mm2 and (d) to (f) have a resolution of mm2. Similar results were obtained for all images in the US liver data set. Overlap measure and computational time are presented in Table 4. The average DCM of the images of the liver data set is with standard deviation of . This performance proofs a good accuracy and high robustness for these challenging imaging. Moreover, as regards our previous method, the number of iterations till convergence and the computational time are reduced significantly, by a factor of 1000 and 15 respectively. A box-plot representation of the DCM is visualized in Figure 6, where the first box corresponds to the US liver data set of and the second one to the data set of . The same accuracy is almost reached for both image resolutions but slightly higher variability is obtained for acquisitions at depth.
(a) DCM values for liver tumors with a depth of 8 (Liver_8) and 16 (Liver_16), eye tumors and prostate; (b) Area overlap (AO) values in prostate segmentation of fP-CMC (using two different types of initial labels: init 1 and init2) vs. GC-FIS .
Fetal head segmentation.
We proceed to define two elliptic labels inside and outside the fetus head (see yellow and cyan contours in Fig. 7) by a mouse click (press, drag and release). Then, the head is segmented with the fP-CMC (red contour in Fig. 7). From there, the axis of elongation of the obtained binary segmentation is computed with the following formulation :(19)where r, c corresponds to the rows and columns in the binary image I (1- for the object and 0 for the background) and , are the coordinates for the centers of mass. The axis of elongation will correspond to Occipito-Frontal Diameter (OFD) measure, and then Biparietal Diameter (BPD) is computed for the angle (), for the same center of mass. Finally, HC is computed from the BPD and OFD measurements using the expression :(20)
fP-CMC segmentation is in red, initial labels (ellipses) are in yellow and light blue for object and background, respectively.
Some visual segmentation results are shown in Figure 7 with their corresponding quantitative measures as in Table 5. Estimated HC metric as compared with radiologist computation for all images is plotted in Figure 8 in form of Bland-Altman plot  where the error in the HC estimation (difference in mm) is represented as a function of the estimated HC measure (in mm). Further, we compare our estimation errors with the inter-user variability of HC measurement reported in previous studies . Sarris et al  recently presented, among other fetal ultrasound measurements, a variability study of HC measurements (in mm) between 3 different users over 175 images from 140 fetuses (ranging from 14 to 41 weeks of gestational age). They reported an average variability of and the confidence intervals at and . Our average error is smaller () and our error distribution fall within those confidence interval. In this sense, we can say that our estimated HC is equivalent to the one measured by hand by the radiologist.
Error measurements of our method w.r.t. the confidence interval at 95% and Bland-Altman plot of inter-subject variability from .
Segmentation of prostate.
In this application we compare our fP-CMC segmentation to previous published works  that are based on graph cut segmentation with a fuzzy interference system (GC-FIS). Both fP-CMC and GC-FIS methodologies are comparable in terms of interactivity and of semi-supervision with labels within a graph-cut based segmentation framework. However, contrary to our philosophy of considering speckle as image information, GC-FIS assumes a preprocessing step for image de-noising in order to reduce the speckle noise by using ‘stick filter’ and smooths the image. The user interaction in GC-FIS differs a lot of ours indeed. In GC-FIS, few points lying on the object contour must be selected first. These points are further used to generate the contours for the inside and the outside labels, being automatically defined in the normal direction and at a certain distance from the object contour. Segmentation is performed next, using the interactive graph-cut approach proposed by Boykov and Jolly in . Please note that GC-FIS is thus based on a discrete optimization approach, while we have a continuous formulation. Finally, the contour is refined in a post processing step using a FIS, where each point is evaluated and marked as weak or strong boundary points. The misclassified (weak) points are used next to add more appropriate hard constraints and to further re-perform the segmentation.
For sake of comparison we force us to define labels as close as possible as defined by GC-FIS. Accordingly, in Fig. 9, for the initial labels shown in the row we obtain the segmentation results as shown in the row. We will denote these results by fP-CMC-L1. Rather than using these complex labels, we will run our algorithm with much simple and intuitive initialization (as performed in all previous clinical applications), by straight lines for the foreground and without background labels. Results are denoted by fP-CMC-L2 and shown in row of Fig. 9.
and rows: initialization labels as in  (yellow for object and blue for background) and fP-CMC-L1 segmentation in red. row: initial labels in yellow and fP-CMC-L2 segmentation in red contour. Ground truth is in transparent green area. We refer to  for visual comparison with GC-FIS results.
In Table 6 we present the quantitative results for both methods in terms of area overlap (AO). Results with fP-CMC-L1 are very close to GC-FIS or slightly better: an overall accuracy average of of fP-CMC-L1 against obtained with GC-FIS. We have considerably reduced the variability of the results, proving again the robustness of the fP-CMC (see boxplots in Fig. 6). Results with fP-CMC-L2, (average accuracy of ) are slightly worse than fP-CMC-L1 and equivalent to GC-FIS. We believe though that this decrease in accuracy is acceptable when considering the much more simple initialization step without background labels. The DCM boxplot for fP-CMC-L2 are shown in Fig. 6.
The most important methodological contributions of our work are the use of a graph of intensity patches for representing the ultrasound image and a fast minimization scheme for the continuous min cut problem. This new formulation has considerably reduced the computation time as regards our previous version . Intuitively, our proposed graph cut method can be seen as a heat diffusion process on the graph of image patches. The heat diffusion propagates the information of the inside and outside labels selected by the user. Regions of interest that wish to be segmented entirely depend on the choice of the labels, i.e. which image patches are selected to be diffused. In other words, our segmentation method can segment both homogeneous and heterogeneous US regions, offering high-quality results with a large flexibility to work on different targets.
In practice, most parameters are easily set. Actually, we found that for many different B-mode sequence and scanners, the same parameter values provided high accurate results (all parameters are equally set for all clinical applications presented). The only parameter that may required an initial fine tuning according to B-mode sequence is the scaling parameter which is more sensitive when the contrast between foreground and background is low, but which can be overcame with a proper label initialization. We showed that despite the semi-supervision, our segmentation framework provides highly accurate and quantitatively similar results for different types of labels and users. Indeed, setting the initial labels (by few clicks, lines, circles, etc.) is extremely easy and fast. The user interaction in the label definition might be seen as a limitation of our method but, in our opinion, ultrasound image analysis is particularly suited for an interactive segmentation framework since, in clinical practice, physicians are used to extract simple biometric measures from few click interactions. Moreover, in cases where ultrasound image quality is really challenging, our framework allows the expert to define labels close to the final object thus providing not only a faster convergence but also more reproducible segmentations. Nevertheless, in an application basis, the label definition process can easily be automatized if needed .
We have evaluated the segmentation results of the fP-CMC framework in several clinical contexts (some of them aimed at quantitative validation purposes more than real clinical applications). We have reported segmentation accuracy in average of for all clinical applications (Fig. 6). This proofs a high practical value of our method in terms of flexibility and easy of use as compared to state of the art methods (limited and optimized for a specific organ and/or image sequence). Our interactive patch-based philosophy is rather different from most existing segmentation techniques for B-mode ultrasound imaging. Nevertheless, we have compared our method to the closest methodology in terms of interactive discrete graph cut segmentation  applied to the prostate.
We have presented a segmentation method for 2D US images since 2D US is widely accepted in clinical practice (except for the eye, all our images where in 2D). However, following the increasing number of 3D US imaging, the extension to 3D images is does not change the mathematics neither the algorithm introduced in this work. Plus, initialization could be easily extended to 3D, for instance considering a few group of 2D slices where the user clicks or to make use of 3D objects (sphere or ellipsoids) as initial labels. We assume that in 3D we would keep the same robustness as in the 2D case. However, this would require confirmation with new experiments, which is out of the scope of the paper.
Finally, we promote here the use of soft priors (semi-supervision with labels) for the ultrasound segmentation. Nevertheless, we agree that the use of strong shape, temporal and/or intensity, priors can be for a specific target more powerful that our soft priors. Note that strong priors can be easily included in our mathematical formulation. Future work goes into exploring other image features to be included in the patch representation (and the corresponding patch distance function) in order to better represent the echogenicity present in US imaging.
The authors would like to thank Prof. Radu Badea and Monica Lupsor, MD-PhD (3rd Medical Clinic, University of Medicine and Pharmacy, Cluj-Napoca, Romania) for providing the liver data set and the manual segmentations too. More, we thank Prof. Francis Munier (Jules Gonin Hospital, Lausanne, Switzerland) that have provided the eye data set along the manual segmentations of the tumors. Our acknowledgment goes also to Prof. Laurent Guibaud and Dr. Benoit Morel (Department of pediatric and fetal imaging, Hôpital Femme Mère Enfant, Lyon, France) that provided the fetal images and biometric measures for the fetal head circumference study. We would like to thank also M. Zouqi (University of Western Ontario, Canada) for his kind cooperation in providing the prostate images. Our acknowledgment also goes to Marcel Arditi, PhD (Bracco Imaging Group SpA, Switzerland) for his thoughtful discussions on ultrasound physics.
Conceived and designed the experiments: AC XB OC NH SN J-PT MBC. Performed the experiments: AC XB OC MBC. Analyzed the data: AC XB OC MBC. Contributed reagents/materials/analysis tools: AC XB OC NH MBC. Wrote the paper: AC XB MBC.
- 1. Corsi C, Saracino G, Sarti A, Lamberti C (2002) Left ventricular volume estimation for real-time three-dimensional echocardiography. IEEE Trans on Medical Imaging 21: 1202–1208.
- 2. Paragios N, Jolly MP, Taron M, Ramaraj R (2005) Active Shape Models and Segmentation of the Left Ventricle in Echocardiography, volume 3459 of Lect. Notes in Computer Science. 131–142 pp.
- 3. Chen Y, Huang F, Tagare HD, Rao M (2007) A coupled minimization problem for medical image segmentation with priors. Int J of Computer Vision 71: 259–272.
- 4. Anquez J, Angelini E, Bloch I (2008) Segmentation of fetal 3D ultrasound based on statistical prior and deformable model. In: IEEE International Symposium on Biomedical Imaging (ISBI). pp. 17–20.
- 5. Chen DR, Chang RF, Kuo WJ, Chen MC, Huang YL (2002) Diagnosis of breast tumors with sonographic texture analysis using wavelet transform and neural networks. Ultrasound Med Biol 28: 1301–1310.
- 6. Unal G, Bucher S, Carlier S, Slabaugh G, Fang T, et al. (2008) Shape-driven segmentation of the arterial wall in intravascular ultrasound images. IEEE Trans Inf Technol Biomed 12: 335–347.
- 7. Noble JA, Boukerroui D (2006) Ultrasound image segmentation: A survey. IEEE Trans on Medical Imaging 25: 987–1010.
- 8. Gong L, Pathak S, Haynor D, Cho P, Kim Y (2004) Parametric shape modeling using deformable superellipses for prostate segmentation. Medical Imaging, IEEE Transactions on 23: 340–349.
- 9. Pingkun Y, Sheng X, Turkbey B, Kruecker J (2011) Adaptively learning local shape statistics for prostate segmentation in ultrasound. Biomedical Engineering, IEEE Transactions on 58: 633–641.
- 10. Chang RF, Wu WJ, Moon WK, Chen WM, Lee W, et al. (2003) Segmentation of breast tumor in three-dimensional ultrasound images using three-dimensional discrete active contour model. Ultrasound in Medicine and Biology 29: 1571–1581.
- 11. Xie J, Jiang Y, tat Tsui H (2005) Segmentation of kidney from ultrasound images based on texture and shape priors. Medical Imaging, IEEE Transactions on 24: 45–57.
- 12. Zouqi M, Samarabandu J (2009) 2D ultrasound image segmentation using graph cuts and local image features. In: Computational Intelligence for Image Processing. pp. 33–40.
- 13. Oosterveld BJ, Thijssen JM, Verhoef WA (1985) Texture in B-mode echograms: 3-D simulations and experiments of the effects of diffraction and scatterer density. Ultrasonic Imag 7: 142–160.
- 14. Tao Z, Tagare HD, Beaty J (2006) Evaluation of four probability distribution models for speckle in clinical cardiac ultrasound images. IEEE Trans on Medical Imaging 11: 1483–1491.
- 15. Efros AA, Leung TK (1999) Texture synthesis by non-parametric sampling. In: ICCV. pp. 1033–1038.
- 16. Buades A, Coll B, Morel JM (2005) A review of image denoising algorithms, with a new one. Multiscale Modeling & Simulation 4: 490–530.
- 17. Coupé P, Hellier P, Kervrann C, Barillot C (2009) Nonlocal means-based speckle filtering for ultrasound images. IEEE Trans Image Process 18: 2221–2229.
- 18. Ledesma-Carbayo MJ, Kybic J, Desco M, Santos A, Suhling M, et al. (2005) Spatio-temporal nonrigid registration for ultrasound cardiac motion estimation. IEEE Trans Image Process 27: 1113–1126.
- 19. Revell J, Mirmehdi M, McNally D (2005) Computer vision elastography: speckle adaptive motion estimation for elastography using ultrasound sequences. IEEE Trans Image Process 24: 755–766.
- 20. Ciurte A, Houhou N, Nedevschi S, Pica A, Munier F, et al.. (2011) An efficient segmentation method for ultrasound images based on a semi-supervised approach and patch-based features. In: IEEE International Symposium on Biomedical Imaging (ISBI). pp. 969–972.
- 21. Boykov Y, Jolly MP (2001) Interactive graph cuts for optimal boundary and region segmentation of objects in n-d images. In: ICCV. volume 1, pp. 105–112.
- 22. Li Y, Sun J, Tang CK, Shum HY (2004) Lazy snapping. In: ACM SIGGRAPH.
- 23. Rother C, Blake A, Kolmogorov V (2004) Grabcut - interactive foreground extraction using iterated graph cuts. In: ACM SIGGRAPH.
- 24. Unger M, Pock T, Cremers D, Bischof H (2008) Tvseg - interactive total variation based image segmentation. In: British Machine Vision Conference (BMVC).
- 25. Houhou N, Bresson X, Szlam A, Chan TF, Thiran JP (2009) Semi-supervised segmentation based on non-local continuous min-cut. In: Scale Space and Variational Methods in Computer Vision. volume 5567, pp. 112–123.
- 26. Chan T, Esedoglu S, Nikolova M (2006) Algorithms for Finding Global Minimizers 575 of Image Segmentation and Denoising Models. SIAM Journal on Applied Mathematics 66(5): 1632–1648.
- 27. Bresson X, Esedoglu S, Vandergheynst P, Thiran J, Osher S (2007) Fast Global Minimization of the Active Contour/Snake Models. Journal of Mathematical Imaging and Vision 28(2): 151–167.
- 28. Chambolle A, Cremers D, Pock T, Bischof H (2009) A Convex Approach for Computing Minimal Partitions. In: IEEE Conference on Computer Vision and Pattern Recognition. pp. 810–817.
- 29. Bae E, Yuan J, Tai XC (2009) Global Minimization for Continuous Multiphase Partitioning Problems Using a Dual Approach. International Journal of Computer Vision 92(1): 112–129.
- 30. Yuan J, Bae E, Tai XC, Boykov Y (2010) A Study on Continuous Max-Flow and Min-Cut Approaches Part I: Binary Labeling. UCLA CAM Report 10–61.
- 31. Lellmann J, Schnörr C (2011) Continuous Multiclass Labeling Approaches and Algorithms. SIAM Journal on Imaging Sciences 4(4): 1049–1096.
- 32. Nieuwenhuis C, Toppe E, Cremers D (2013) A Survey and Comparison of Discrete and Continuous Multi-label Optimization Approaches for the Potts Model. International Journal of Computer Vision.
- 33. Strang G (1983) Maximal flow through a domain. Mathematical Programming 26: 123–143.
- 34. Cheng J, Foo SW, Krishnan S (2006) Watershed-presegmented snake for boundary detection and tracking of left ventricle in echocardiographic images. Information Technology in Biomedicine, IEEE Transactions on 10: 414–416.
- 35. Mignotte M, Meunier J, Tardif JC (2001) Endocardial boundary etimation and tracking in echocardiographic images using deformable template and markov random fields. Pattern Analysis and Applications 4: 256–271.
- 36. Chiang HH, Cheng JZ, Hung PK, Liu CY, Chung CH, et al.. (2010) Cell-based graph cut for segmentation of 2D/3D sonographic breast images. In: IEEE International Symposium on Biomedical Imaging. pp. 177–180.
- 37. Aleman-Flores M, Alvarez L, Caselles V (2008) Breast nodule ultrasound segmentation through texture-based active contours. In: Progress in Industrial Mathematics at ECMI 2006, Springer Berlin Heidelberg, volume 12 of Mathematics in Industry. pp. 858–862.
- 38. Xiao G, Brady M, Noble J, Yongyue Z (2002) Segmentation of ultrasound b-mode images with intensity inhomogeneity correction. Medical Imaging, IEEE Transactions on 21: 48–57.
- 39. Shen D, Zhan Y, Davatzikos C (2003) Segmentation of prostate boundaries from ultrasound images using statistical shape model. Medical Imaging, IEEE Transactions on 22: 539–551.
- 40. Shao F, Ling K, Ng W, Wu R (2003) Prostate boundary detection from ultrasonographic images. J Ultrasound Med 22: 605–623.
- 41. Cvancarova M, Albregtsen F, Brabrand K, Samset E (2005) Segmentation of ultrasound images of liver tumors applying snake algorithms and gvf. International Congress Series 1281: 218–223.
- 42. Lee WL, Chen YC, Chen YC, Hsieh KS (2005) Unsupervised segmentation of ultrasonic liver images by multiresolution fractal feature vector. Information Sciences 175: 177–199.
- 43. Huang X, Dione D, Compas C, Papademetris X, Lin B, et al.. (2012) A dynamical appearance model based on multiscale sparse representation: Segmentation of the left ventricle from 4d echocardiography. In: Ayache N, Delingette H, Golland P, Mori K, editors, Medical Image Computing and Computer-Assisted Intervention MICCAI 2012, Springer Berlin Heidelberg, volume 7512 of Lecture Notes in Computer Science. pp 58–65.
- 44. Huang X, Dione DP, Compas CB, Papademetris X, Lin BA, et al. (2014) Contour tracking in echocardiographic sequences via sparse representation and dictionary learning. Medical Image Analysis 18: 253–271.
- 45. Loupas T, McDicken WN, Allan PL (1989) An adaptive weighted median filter for speckle suppression in medical ultrasonic images. IEEE Transactions on Circuits and Systems 36: 129–135.
- 46. Boykov Y, Kolmogorov V (2004) An experimental comparison of min-cut/max-ow algorithms for energy minimization in computer vision. EEE Transactions on Pattern Analysis and Machine Intelligence 26: 1124–1137.
- 47. Boykov Y, Funka-Lea G (2006) Graph Cuts and Efficient ND Image Segmentation. International Journal of Computer Vision 70(2): 109–131.
- 48. Shi J, Malik J (2000) Normalized cuts and image segmentation. Pattern Analysis and Machine Intelligence 22: 888–905.
- 49. Glowinski R, Marrocco A (1975) Sur l'Approximation, par Eléments Finis d'Ordre Un, et la Résolution, par Pénalisation-Dualité, d'une Classe de Problèmes de Dirichlet Non Linéaires. Laboria.
- 50. Glowinski R, Tallec PL (1989) Augmented Lagrangian and Operator-Splitting Methods in Nonlinear Mechanics. SIAM.
- 51. Bioucas-Dias J, Figueiredo M (2007) A new TwIST: Two-step iterative shrinkage/thresholding algorithms for image restoration. Image Processing, IEEE Transactions on 16: 2992–3004.
- 52. Goldstein T, Osher S (2009) The split bregman method for L1-regularized problems. SIAM J Img Sci 2: 323–343.
- 53. Setzer S (2011) Operator splittings, bregman methods and frame shrinkage in image processing. Int J Comput Vision 92: 265–280.
- 54. Donoho D (1995) De-noising by soft-thresholding. Information Theory, IEEE Transactions on 41: 613–627.
- 55. Duval V, Aujol JF, Gousseau Y (2010). On the parameter choice for the non-local means.
- 56. Lee A, Pedersen K, Mumford D (2003) The nonlinear statistics of high-contrast patches in natural images. International Journal of Computer Vision 54: 83–103.
- 57. Fledelius HC (1997) Ultrasound in ophthalmology. Ultrasound in Medicine and Biology 23: 365–375.
- 58. Fenster A, Downey D, Cardinal H (2001) Three-dimensional ultrasound imaging. Phys Med Biol 46: 67–99.
- 59. Cuadra MB, Gorthi S, Karahanoglu FI, Salvador F, Pica A, et al.. (2009) Model-based segmentation and fusion of 3d computed tomography and 3d ultrasound of the eye for radiotherapy planning. In: Computational Vision and Medical Image Processing. pp. 53–58.
- 60. Rueda S, Fathima S, Knight C, Yaqub M, Papageorghiou A, et al.. (2013) Evaluation and comparison of current fetal ultrasound image segmentation methods for biometric measurements: A grand challenge. Medical Imaging, IEEE Transactions on.
- 61. Dice LR (1945) Measures of the amount of ecologic association between species. Ecology 26: 297–302.
- 62. Grady L (2006) Random walks for image segmentation. Pattern Analysis and Machine Intelligence, IEEE Transactions on 28: 1768–1783.
- 63. Flusser J (2000) On the independence of rotation moment invariants. Pattern Recognition 33: 1405–1410.
- 64. Altman D, Chitty L (1997) New charts for ultrasound dating of pregnancy. Ultrasound Obstet Gynecol 10: 174–191.
- 65. Sarris I, Ioannou C, Chamberlain P, Ohuma E, Roseman F, et al. (2012) Intra- and interobserver variability in fetal ultrasound measurements. Ultrasound in Obstetrics and Gynecology 39: 266–273.