Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Multi Texture Analysis of Colorectal Cancer Continuum Using Multispectral Imagery

  • Ahmad Chaddad ,

    Affiliations Laboratory for Imagery, Vision and Artificial Intelligence, École de Technologie Supérieure, Montréal, Québec, Canada, Laboratory of Conception, Optimization and Modelling of Systems, University of Lorraine, Metz, Lorraine, France

  • Christian Desrosiers,

    Affiliation Laboratory for Imagery, Vision and Artificial Intelligence, École de Technologie Supérieure, Montréal, Québec, Canada

  • Ahmed Bouridane,

    Affiliation School of Computing, Engineering and Information Sciences, Northumbria University, Newcastle, United Kingdom

  • Matthew Toews,

    Affiliation Laboratory for Imagery, Vision and Artificial Intelligence, École de Technologie Supérieure, Montréal, Québec, Canada

  • Lama Hassan,

    Affiliation Laboratory of Conception, Optimization and Modelling of Systems, University of Lorraine, Metz, Lorraine, France

  • Camel Tanougast

    Affiliation Laboratory of Conception, Optimization and Modelling of Systems, University of Lorraine, Metz, Lorraine, France



This paper proposes to characterize the continuum of colorectal cancer (CRC) using multiple texture features extracted from multispectral optical microscopy images. Three types of pathological tissues (PT) are considered: benign hyperplasia, intraepithelial neoplasia and carcinoma.

Materials and Methods

In the proposed approach, the region of interest containing PT is first extracted from multispectral images using active contour segmentation. This region is then encoded using texture features based on the Laplacian-of-Gaussian (LoG) filter, discrete wavelets (DW) and gray level co-occurrence matrices (GLCM). To assess the significance of textural differences between PT types, a statistical analysis based on the Kruskal-Wallis test is performed. The usefulness of texture features is then evaluated quantitatively in terms of their ability to predict PT types using various classifier models.


Preliminary results show significant texture differences between PT types, for all texture features (p-value < 0.01). Individually, GLCM texture features outperform LoG and DW features in terms of PT type prediction. However, a higher performance can be achieved by combining all texture features, resulting in a mean classification accuracy of 98.92%, sensitivity of 98.12%, and specificity of 99.67%.


These results demonstrate the efficiency and effectiveness of combining multiple texture features for characterizing the continuum of CRC and discriminating between pathological tissues in multispectral images.


Colorectal cancer (CRC) is a common malignancy having an increasing incidence in many developed countries. It is the third most common newly diagnosed cancer, accounting for 8% of new cases each year, and also the third most common cause of cancer death in both men and women [1]. An estimated 26,270 men and 24,040 women died of colorectal carcinoma in 2014, as reported by the American Cancer Society. Surgical resection of the primary tumor with curative intent is possible in only 70% of patients [2, 3]. Unfortunately, up to 30% of CRC patients who undergo surgical resection of the primary tumor experience a subsequent relapse within three years, with a median time to death of 12 months [4]. Imaging studies are frequently used to evaluate patients for the screening and staging of colorectal cancer. Cross sectional imaging techniques like computed tomography (CT) [4], magnetic resonance imaging (MRI) [5] and microscopy provide anatomical and morphological information about the structure and patterns of the tumor [6]. In this information, image texture, and in particular texture heterogeneity, is a prominent feature of CRC that manifests itself as areas of high cell density.

A recent computational trend has been the analysis of high-resolution whole slide images produced from digital pathology slides [7, 8]. Texture features extracted from such images serve as an input to important computational applications such as computer-aided diagnosis from pathology. An early study by Esgiar et al. showed that entropy texture features extracted from gray-level co-occurrence matrices (GLCM) were capable of differentiating between normal and cancerous tissue [6]. A follow up study by the same authors incorporated fractal dimensions into the feature analysis to improve the sensitivity and specificity of classification [9]. Using color channel histograms, GLCM and structural features, Kalkan et al. achieved an accuracy of 75.15% in the classification of four types of colon tissues: normal, cancerous, adenomatous and inflammatory [10]. Jiao et al. proposed a method for automatic colon cancer detection, using GLCM for texture extraction and support vector machines (SVM) for classification. This method achieved an accuracy of 96.67% in differentiating between cancerous and non-cancerous images [11]. Hilado et al. used 2D discrete wavelet (DW) transform features to classify whole slide colon cancer images into normal, cancerous and adenomatous polyp cases, reporting a 91.11% accuracy [12]. Francesca et al. used whole tumor texture features, computed using Laplacian-of-Gaussian (LoG) filters to assess the heterogeneity of CRC [4]. In a later study, Rao et al. considered LoG texture features to discriminate between CRC patients with and without hepatic metastases [13]. Various approaches using local descriptors have also been proposed, including methods based on scale-invariant feature transform (SIFT) [14], shape context [15], and histograms of oriented gradient (HOG) descriptors [16]. Because they rely on key points that may vary from one subject to another, key point-based SIFT and shape context features methods are not always suitable for the assessment of cellular abnormalities from optical microscopy systems. Likewise, methods based on HOG are not invariant to rotations that occur in microscopic images. There are thus strong arguments supporting the use of rotation-invariant texture features derived from GLCM, LoG filters and DW for this specific problem.

As a general methodology, pathological tissues (PT) resulting from cellular abnormalities in CRC, such as benign hyperplasia (BH), intraepithelial neoplasia (IN) and carcinoma (Ca), can be detected from classical optical microscopy systems using a variety of image processing techniques [1719]. This study proposes to model the continuum of CRC using rich, informative texture features obtained from multispectral optical microscopy images. The discriminative ability of texture features can be appreciated from Fig 1, where the histogram of pixel intensities is shown for images of the BH, IN and Ca types. It can be seen that BH, IN and Ca tissues exhibit noticeably different intensity profiles, supporting the idea that such feature can be used to differentiate between these PT types. The novelty of this work lies in the comparative analysis and combination of three different texture features based on GLCM [17], LoG [4] and DW [20], for predicting PT types. As will be shown in our experiments, using multi-textural information can improve the detection and classification of pathological tissues, and provide a more comprehensive understanding of the connection between CRC and tissue heterogeneity. The potential impacts of this work on the improvement of medical care are two-fold. In combination with standard screening approaches for CRC, the proposed method could improve the detection of the disease in its early stages, thereby increasing the chances of successful treatment. According to the American Cancer Society, the 5-year relative survival rate is about 90% when CRC can be detected before it has spread [21]. The classification of cellular abnormalities in pathological tissues is also essential to assess the progression of CRC and select the appropriate course of treatment. By using texture features, our method provides an effective way of characterizing tissue properties at the cellular level. Measuring the textural information at various points in time could help track the disease’s progression and evaluate the efficiency of a given treatment.

Fig 1. Images of three different types of pathological tissues.

(a) Benign Hyperplasia; (b) Intraepithelial Neoplasia; (c) Carcinoma; (d, e, and f) Histograms show pixel intensity distributions for each type.

Materials and Methods

This study was approved by the institutional review board of the Anatomical pathology (Anapath) Department at the CHU Nancy-Brabois Hospital. The review board waived the need for written informed consent from the participants. Part of the data was used in previous studies [18, 22, 23].

The proposed framework, shown in Fig 2, consists in a series of five steps: 1) sample preparation and image acquisition, 2) ROI segmentation, 3) texture feature extraction, 4) PT type classification, and 5) performance evaluation. A detailed presentation of each step is given in the following sections.

Fig 2. Image analysis processing pipeline: multispectral image acquisition, contour-based segmentation of abnormal tissues, texture feature extraction and finally classification.

(a) Optical microscopy system, staining, sectioning, and scanning. (b) Multispectral image acquisition via a CCD camera across a range of visual spectral bands. (c) Active contour segmentation algorithm for delineating ROIs. (d) GLCM, LoG and DW image texture feature extraction. (e) Supervised classification for automatic prediction of abnormal tissue types from new samples.

Sample preparation and image acquisition

CRC data were collected from the Anatomical pathology (Anapath) Department at the CHU Nancy-Brabois Hospital. Tissue samples were obtained from sequential resections of colons from 30 patients with colorectal cancer. Sections of 5 μm thickness were extracted and stained with the widely used haematoxylin and eosin (H&E) stains, to reduce image processing requirements. Images were captured at low magnification (× 40) using an optical microscope mounted with a charge coupled device (CCD) camera. A Liquid Crystal Tunable Filter (LCTF) was placed in the optical path between the light source and the CCD camera, providing a bandwidth accuracy of 5 nm and a controllable wavelength through the visible spectrum from 500 to 650 nm [24]. This technique, known as hyperspectral or multispectral imaging [25], can capture images of tissue samples at different spectral frequencies. In this study, 16 multispectral bands were acquired in the wavelength range of 500–650 nm, with 9.375 nm steps between successive bands.

The data of 30 CRC patients (10 BH patients, 9 IN patients and 11 Ca patients) were obtained for analysis, giving a total of 160 BH images, 144 IN images, and 176 Ca images. Data are available in S1 Zip ( or DOI: 10.6084/m9.figshare.2076220). The identification of each PT type from multispectral images was done by a senior histopathologist, confirming the diagnosis. Finally, images were denoised using an average filter and rescaled to a resolution of 512 × 512 pixels.

Active contour based segmentation

Active contour segmentation was used to identify tissue boundaries within the image. This segmentation technique, which moves a dynamic curve iteratively towards object contours in the image, is well adapted to delineate irregular shapes [26, 27]. While it can achieve a high accuracy, it can also suffer from long runtimes. To accelerate the segmentation process, we limited the number of iterations based on empirical calculations [17]. A multi-scale approach was used to further reduce runtimes by performing an initial segmentation at a coarse resolution of 64 × 64 pixels and then refining this solution at a high resolution of 512 × 512 pixels. Using this technique, segmentations were obtained in less than a minute with a standard PC (Intel Core i5 3.4 GHz processor with 32 GB RAM). Note that the runtime performance can be improved via alternative specialized computational technologies, such as a pipeline algorithm based on field programmable gate array (FPGA) technology [28].

Ground truth segmentations, one segmentation per sample, were obtained manually using 3D Slicer [29] and validated by two pathologists. An example of a ground truth segmentation is given in Fig 3. The ground truth images were used to evaluate the performance of the active contour segmentation (Fig 4), based on the following metrics: Dice similarity coefficient (DSC), false positive rate (FPR) and false negative rate (FNR). DSC measures the degree of the correspondence (similarity) between ground truth and segmented ROIs, and is defined as (1) where A and B are pixel sets corresponding to ground truth and segmented regions, respectively.

Fig 3. Examples of ground truth segmentation.

(a) Original image. (b) Segmented image. (c) Labeled image. Labeled area in (c) corresponds to the ROI used for texture feature extraction.

Fig 4. Examples of active contour segmentation.

Top-row images correspond to the (a) BH, (b) IN and (c) Ca types. Bottom-row images (aˊ), (bˊ), and (cˊ) show the ROI obtained on these images by the active contour segmentation method.

The FPR and FNR are two measures of over and under segmentation, and are calculated as follows: (2) (3) where contains the pixels that are not in the ground truth set A.

Texture feature extraction

Three types of texture filters, based on LoG, DW and GLCM, were applied to the segmented ROIs in each spectral band. Texture features were then extracted from the filtered images by applying specialized quantifier functions (Fig 2).

LoG based Texture.

The Laplacian of Gaussian (LoG) filter can be viewed as the combination of a Gaussian smoothing operator with a kernel of width sigma (σ) and an isotropic filter, the Laplacian, which measures the second spatial derivative in the image. The LoG is commonly used to detect edges and blobs at various scales. For each segmented ROI, a LoG filter was applied using σ values of 0.5 (fine texture type), 1.5 (medium texture type) and 2.5 (coarse texture type). The region was then quantified by computing the Average (A), Standard deviation (SD) and Entropy (Ent) of its LoG values. The theoretical framework of this technique is described in detail in the Appendix B of Ganeshan et al [30]. Let f(x,y) be the LoG value of a pixel (x,y) in the segmented region Ω. For entropy calculation, we discretize the distribution of LoG values into 256 equal-sized intervals and denote as Ωk the subset of pixels within the k-th interval. The LoG quantifier functions can be defined as follows: (4) (5) (6)

For each sample, we thus obtained a set of 9 texture: (7) where f, m, and c represent the fine, medium and coarse texture.

Texture based on Discrete Wavelet (DW) Transform.

The DW Transform (DWT) analyzes an image by decomposing it into a coarse approximation via low-pass filtering and a detailed component via high-pass filtering. The decomposition is performed recursively on the low-pass approximation coefficients obtained at each level, until the necessary number of iterations is reached [31]. Four decomposition directions (sub-bands) are considered: horizontal (0°, Dh), first diagonal (45°, Dd), vertical (90°, Dv) and second diagonal (135°, Dd). The decomposition at the each level i provides an approximation matrix (image) Ai and three detail matrices, namely, Dhi (horizontal coefficient matrix), Dvi (vertical coefficient matrix) and Ddi (diagonal coefficient matrix).

A one level 2D DWT decomposition was applied on each ROI, encoding the region’s texture as a low-frequency component A1 and three high-frequency components: Dd1 (diagonal detail), Dv1 (vertical detail), and Dh1 (horizontal detail) (Fig 5). Three quantifier functions, measuring entropy (f1DW), energy (f2DW) and variance (f3DW), were applied to the average DWT coefficients matrix (i.e., the average of A1, Dh1, Dv1 and Dd1 matrices) of Daubechies (db), Coiflet (coif) and Symlet (sym) filters, respectively. For each sample, 9 DW-based texture features were obtained, corresponding to the following feature vector: (8)

Fig 5. Image DW transform decomposition.

R corresponds to rows, C corresponds to columns, l and h are the index of low and high pass filter respectively, 2ds1 and 1ds2 are the down-sample columns and rows respectively, and {×} is the convolution operator.

Texture based gray level co-occurrence matrix.

Proposed by Haralick in [32], the gray-level co-occurrence matrix (GLCM) is an efficient texture analysis method which uses second-order statistics to characterize the properties of two or more pixel values occurring at specific locations relative to each other. Formally, GLCM matrices represent the probabilities Pd,θ (i, j) of transition from a pixel with intensity i to a pixel of intensity j separated by a translation vector defined by direction θ and offset d (also known as distance). Given a 2D image I of size N × N, the co-occurrence matrix Pd,θ (i, j) can be defined as (9) where dx and dy specify the distance between the pixel of interest and its neighbor along the x-axis and the y-axis of an image, respectively. The GLCM matrix has a size of Ng × Ng, where Ng is the number of gray levels in the image.

We extracted the GLCM features as follows. A histogram equalization was first applied on the segmented ROIs encoded using 256 grey levels. GLCM matrices were then calculated based on four offsets (1, 2, 3 and 4 pixels) and four phases (0°, 45°, 90°, 135°). Using this technique, we obtained 16 GLCM matrices of size 256 × 256 for each segmented ROI. Texture features were then calculated by applying 12 quantifier functions on each GLCM matrix, and averaging the values across the 16 GLCM matrices. The 12 quantifier functions, proposed by Haralick et al. [32] and Clausi D. [33], are reported. Texture features based on the quantification of GLCM values can be expressed as the following vector: (10)

Statistical analyses and classification criteria.

A statistical analysis based on the Kruskal-Wallis test was carried out to measure the significance of texture features for discriminating between the three PT types. Features having a p-value of 0.01 or less were considered as statistically significant. Note that this analysis is not used for feature selection and, consequently, does not bias the following classification step.

Four classification methods were tested: linear discriminant analysis (LDA) [34], naïve Bayes (NB) [35], decision trees (DT) [36] and nearest neighbors (NN) classifier [37]. In LDA, the probability distribution functions (PDF) of classes are assumed to be multivariate Gaussian with different mean but same covariance matrix, and Bayes classification is used to select the class with maximum probability for each test sample. Naïve Bayes considers features as independent from one another, given their class, and computes the PDF parameters of these features for each class. A univariate Gaussian PDF is assumed for all features. DT classification splits the set of training samples recursively, by applying a threshold on selected features, until all leaf nodes are sufficiently pure (i.e., they contain samples of the same class) or a maximum number of levels is attained. In this work, the Gini index was used as measure of purity [38]. NN finds the K training samples nearest to a given test sample, based on the Euclidean distance, and assigns the test sample to the most frequent class of its nearest neighbors. Based on prior testing, a value of K = 10 was used for the number of nearest neighbors.

Classifier performance is evaluated using three metrics: accuracy, sensitivity, and specificity [39]. Accuracy measures the proportion of test samples correctly classified by the method. Sensitivity computes, for each class, the ratio between the number of true positives (i.e., positives samples classified as positive by the method) and total number of positive samples. The values obtained for each class are then averaged proportionally to the number of samples in the corresponding class. Likewise, specificity is the proportion of negative samples that are classified as negative by the method. In addition, the area under the receiver operating characteristic curve (AUC) is used to evaluate the classifiers’ performance for different decision thresholds. AUC values are obtained by plotting the curve of true positive rate (i.e., sensitivity) versus false positive rate (i.e., 1-specificity), for various decision thresholds, and measuring the total area under the curve. A higher AUC value indicates a better classifier.

A 10-fold cross-validation approach was employed to obtain unbiased estimates of classifier performance. In this approach, the data is first partitioned into 10 equal-sized sample sets. Each set is then held-out in turn for validation, while the remaining samples are used for training [40]. The average performance, computed over these 10 folds, is reported.



Table 1 shows the segmentation accuracy in terms of DSC, FPR and FNR obtained by the active contour segmentation method, for images of the three PT types. We observe DSC values in the range of 86.31%–88.21%, with the best performance achieved for Ca regions. Moreover, ranges of 5.03%–7.61% and 16.11%–20.26% were obtained for FPR and FNR, respectively, the lowest values corresponding to IN regions (FPR = 5.03%) and Ca regions (FNR = 16.11%). These results confirm the ability of active contour segmentation to accurately extract the ROIs, in particular, regions corresponding to Ca.

Table 1. Average performance metrics (%) for the three pathological tissue types.

Texture analysis

The mean and standard deviation of LoG-based texture features, obtained at different filter scales, are shown in Table 2. A Kruskall-Wallis test was used for each feature to determine if its distribution of values differs across PT types. In this test, the null hypothesis is that the mean rank of values is the same for each type. The p-values obtained for each feature are given in the last column of Table 2. Except for Ac, the mean rank of LoG features is significantly higher in Ca than IN and BH (p-value < 0.0001). Likewise, the mean rank of IN is higher than BH, except for features Ac and Entf. This supports the idea that PT types have different textural properties and that features based on LoG can be used to discriminate between these types of tissue abnormality.

Table 2. Mean (± standard deviation) of LoG texture features at different PT types.

A similar analysis was conducted for GLCM features (Table 3) and DW features (Table 4). For GLCM, the mean rank of features f1, f2, f4, f8, f11 and f12 is significantly higher in Ca than IN and BH types. In contrast, the mean rank of feature f9 is higher in IN compared to BH and Ca. Additionally, the mean rank of features f5, f6, f7 and f10 was found to be significantly higher in BH than IN (p-value < 0.0001). For features extracted using the DW transform, we found that features f1DW_db, and f1DW_sym had a higher mean rank for IN than BH and Ca types, and that the mean rank of features f1DW_coif, f2DW_db, f2DW_coif, f2DW_sym, f3DW_db, f3DW_coif and f3DW_sym was higher in BH than IN and Ca types (p-value < 0.001).

Table 3. Mean (± standard deviation) of texture features extracted from GLCM of the different PT types.

Table 4. Mean (± standard deviation) of DW based feature at the different PT types.

Overall, the analysis shows the potential of LoG, GLCM and DW texture features for differentiating between PT types. Furthermore, since all texture features were found to be statistically significant (p-value < 0.01), all of them were used for classification (i.e., no feature selection step was performed prior to classification).


The three sets of texture features (i.e., FLOG, FDW and FGLCM) were evaluated in a classification setting, using them as input to LDA, NB, DT and NN classifiers. The performance, in terms of accuracy, sensitivity and specificity, obtained by these classifiers for each feature set is reported in Table 5. The best classification accuracies obtained for LoG, DW and GLCM are 81.17% (DT classifier), 90.00% (LDA classifier) and 94.37% (LDA classifier), respectively.

Table 5. Summary of image classification: 160 BH, 144 IN and 176 Ca.

The ability of the classifiers to discriminate between pairs of PT types, for various decision thresholds, was evaluated using the AUC metric. Results, shown in Table 6 and Fig 6, indicate that all three sets of texture features are useful to discriminate between all pairs of PT types, with AUC values ranging from 98% to 100%. Moreover, Table 7 gives the confusion matrix obtained for the three types of texture features. We see that, for all texture types, the highest accuracy is achieved for Ca (167/176 correctly classified Ca samples using GLCM), and the most frequent classification error occurs between BH and IN types.

Fig 6. ROC analysis for the assessment of continuum PT.

The blue, black, and red line is for BH vs. IN, BH vs. Ca, and IN vs. Ca respectively. (a) Texture based on LoG filter. (b) Texture based on DWT filter. (c) Texture based on GLCM.

To further improve the classification performance, we concatenated the features derived from all three types of texture, giving a vector of 30 texture features. Using this approach, we obtained an accuracy of 98.92%, a sensitivity of 98.12%, a specificity of 99.67%, and an AUC of 100% using the LDA classifier (Tables 5 and 6). The performance improvement obtained with multi-texture features can also be observed in the confusion matrix of Table 7, with 157/160 of BH, 143/144 of IN, and 175/176 of Ca samples correctly classified.

Randomization test

Randomized permutation tests were used to further quantify the significance of the link between image textures and PT types. Multiple trials were conducted to compute the classification accuracy from randomly permuted PT type labels. This approach allows the quantification of the null distribution of texture feature classification accuracy, i.e. given the null hypothesis that features contain no information regarding PT types, see permutation testing [41]. The analysis was performed as before, except that type labels were randomly permuted prior to evaluation, thereby generating an empirical null distribution over classification results from multiple trials (1000 times). As expected, the null distribution is peaked around classification values equivalent to random guessing. e.g. accuracy = 33.25% (median = 32.86%) for texture features based on LoG filter, 33.81% (median = 33.75%) for features based on DW, 33.83% (median = 33.95%) for the features derived from GLCM, and 33.91% (median = 33.93%) for full feature set (combined features), (Table 8). These distributions can be used to calculate empirical p-values for classification results obtained in experiments in Section 3, e.g. Table 5, which are in the significant range.


A multispectral image processing pipeline was presented, in which regions of interest (ROIs) representing abnormal tissues are automatically segmented via an efficient multi-resolution active contour method. This method was shown to be accurate, with respect to an expert labeled ground truth, obtaining Dice similarity values between 86.31% and 88.21%.

In a comparative study, we evaluated the usefulness of three types of textures for classifying pathological tissues related to CRC. Individually, all textures lead to classification accuracies above 80%, although GLCM based textures provided the best performance with an accuracy of 94.37%, sensitivity of 95.63% and specificity of 100% (Table 5). Comparing the performance across PT types, we observed that Ca samples have the lowest error rate, and that most errors occurred between BH and IN types (Table 7). We also observed that combining all three texture types (for a total of 30 features) provides the best performance, with an accuracy of 98.92%, sensitivity of 98.12%, specificity of 99.67% and AUC of 100% (Tables 5, 6 and 7). This indicates that different texture features contain complementary information, which can be combined in a synergistic fashion to improve the analysis.

Fig 7 shows the correlation between the feature values found in the three PT types, for each type of texture. For LoG based texture features, the highest correlation values are observed between the medium and coarse textures of IN and Ca types (Fig 7A). A similar correlation pattern is found between the wavelet features within the IN and Ca types (Fig 7B). In contrast, less correlation is observed between different GLCM features (Fig 7C), which could explain their relatively high classification accuracy. Across PT types, low correlation values are observed between features, especially for LoG and GLCM textures. Once more, this supports the driving hypothesis that texture features can be used to characterize and identify pathological tissues in multispectral CRC images.

Fig 7. Heatmap of correlation coefficients between PT features: BH, IN, and Ca is the begin hyperplasia, intraepithelial neoplasia and carcinoma respectively.

(a) Texture feature based on LoG filter, A, Ent and SD is the Average, Entropy and Standard Deviation respectively; f, m and c is the index of fine, medium and coarse texture respectively. (b) Texture feature based on discrete wavelet where f1, f2, and f3 is the Entropy, Energy and Variance respectively; d, c and s is the index of Daubechies, Coiflet and Symlet wavelet respectively. (c) Texture feature extraction from GLCM where f1, f2, f3, f4, f5, f6, f7, f8, f9, f10, f11 and f12 is the index of Energy, Entropy, Correlation, Contrast, Inverse difference, Sum-variance, Sum-mean, Difference entropy, Cluster shade, Cluster tendency, Maximum probability, and Difference variance respectively.

Finally, we note that several studies in the literature have outlined the advantages of using texture features to identify abnormal colon samples [6, 9, 10, 42, 43, 44]. However, few of them have focused on distinguishing cancer grades using the progression of pathological tissues. One such study uses multiscale local binary patterns (LBP) and support vector machines (SVM) to analyze colorectal tumor biopsies, reporting a classification accuracy of 91.28% [22]. Deep learning methods have also been used for both classification and feature learning in colon histopathology images, achieving an accuracy of 97.30% for classifying between cancer and non-cancer images [45]. While such studies did not use multispectral images nor considered the continuum of CRC (i.e., various PT types), as in our work, deep learning methods like convolution neural networks (CNN) show a great promise in improving the analysis of CRC. In particular, these methods would allow the analysis of texture at various image scales.


Multispectral texture analysis is a promising noninvasive approach to quantify the spatial heterogeneity of pathological tissues (PT) associated with CRC. Our results suggest that such analysis can help detect PT in the progression of CRC from benign cell proliferation to malignant lesions. Future work will further investigate the relationship between image texture features and CRC. In particular, using a larger variety of tissue samples and cellular abnormalities would help evaluate the generalizability of the proposed method, possibly in the context of other types of cancer. Moreover, since texture features can be acquired from arbitrary imaging modalities, the usefulness of our texture-based analysis approach could also be tested on modalities other than optical microscopy, such as MRI or CT.

Supporting Information

S1 Zip. Multispectral images.

BH: Begin Hyperplasia; IN: Intraepithelial Neoplasia; Ca: Carcinoma.



The authors would like to thank the Anatomical pathology (Anapath) Department of the CHU Nancy-Brabois Hospital for providing the data used in this study.

Author Contributions

Conceived and designed the experiments: AC. Performed the experiments: AC. Analyzed the data: AC CD CT MT. Contributed reagents/materials/analysis tools: AC CT. Wrote the paper: AC CD AB MT LH CT.


  1. 1. Siegel R, DeSantis C, Jemal A. Colorectal cancer statistics, 2014. CA Cancer J Clin. 2014;64: 104–117. pmid:24639052
  2. 2. Sargent DJ, Wieand HS, Haller DG, Gray R, Benedetti JK, Buyse M, et al. Disease-free survival versus overall survival as a primary end point for adjuvant colon cancer studies: individual patient data from 20,898 patients on 18 randomized trials. J Clin Oncol Off J Am Soc Clin Oncol. 2005;23: 8664–8670.
  3. 3. Sargent DJ, Patiyil S, Yothers G, Haller DG, Gray R, Benedetti J, et al. End points for colon cancer adjuvant trials: observations and recommendations based on individual patient data from 20,898 patients enrolled onto 18 randomized trials from the ACCENT Group. J Clin Oncol. 2007;25: 4569–4574. pmid:17876008
  4. 4. Ng F, Ganeshan B, Kozarski R, Miles KA, Goh V. Assessment of Primary Colorectal Cancer Heterogeneity by Using Whole-Tumor Texture Analysis: Contrast-enhanced CT Texture as a Biomarker of 5-year Survival. Radiology. 2013;266: 177–184. pmid:23151829
  5. 5. Low RN. MRI of colorectal cancer. Abdom Imaging. 2002;27: 418–424. pmid:12066240
  6. 6. Esgiar AN, Naguib RN, Sharif BS, Bennett MK, Murray A. Microscopic image analysis for quantitative measurement and feature identification of normal and cancerous colonic mucosa. IEEE Trans Inf Technol Biomed Publ IEEE Eng Med Biol Soc. 1998;2: 197–203.
  7. 7. Cooper LAD, Kong J, Gutman DA, Dunn WD, Nalisnik M, Brat DJ. Novel genotype-phenotype associations in human cancers enabled by advanced molecular platforms and computational analysis of whole slide images. Lab Investig J Tech Methods Pathol. 2015;95: 366–376.
  8. 8. Loughrey MB, Kelly PJ, Houghton OP, Coleman HG, Houghton JP, Carson A, et al. Digital slide viewing for primary reporting in gastrointestinal pathology: a validation study. Virchows Arch Int J Pathol. 2015;467: 137–144.
  9. 9. Esgiar AN, Naguib RNG, Sharif BS, Bennett MK, Murray A. Fractal analysis in the detection of colonic cancer images. IEEE Trans Inf Technol Biomed. 2002;6: 54–58. pmid:11936597
  10. 10. Kalkan H, Nap M, Duin RPW, Loog M. Automated classification of local patches in colon histopathology. 2012 21st International Conference on Pattern Recognition (ICPR). 2012. pp. 61–64.
  11. 11. Jiao L, Chen Q, Li S, Xu Y. Colon Cancer Detection Using Whole Slide Histopathological Images. In: Long M, editor. World Congress on Medical Physics and Biomedical Engineering May 26–31, 2012, Beijing, China. Springer Berlin Heidelberg; 2013. pp. 1283–1286. Available:
  12. 12. Hilado SDF, Gan Lim LA, Naguib RNG. Implementation of Wavelets and Artificial Neural Networks in Colonic Histopathological Classification. J Adv Comput Intell Intell Inform. 2014;18: 792–797.
  13. 13. Rao S-X, Lambregts DM, Schnerr RS, van Ommen W, van Nijnatten TJ, Martens MH, et al. Whole-liver CT texture analysis in colorectal cancer: Does the presence of liver metastases affect the texture of the remaining liver? United Eur Gastroenterol J. 2014;2: 530–538.
  14. 14. Lowe DG. Distinctive Image Features from Scale-Invariant Keypoints. Int J Comput Vis. 2004;60: 91–110.
  15. 15. Belongie S, Malik J, Puzicha J. Shape matching and object recognition using shape contexts. IEEE Trans Pattern Anal Mach Intell. 2002;24: 509–522.
  16. 16. Dalal N, Triggs B. Histograms of Oriented Gradients for Human Detection. In CVPR. 2005. pp. 886–893.
  17. 17. Chaddad A, Tanougast C, Dandache A, Bouridane A. Extracted haralick’s texture features and morphological parameters from segmented multispectrale texture bio-images for classification of colon cancer cells. WSEAS Trans Biol Biomed. 2011;8: 39–50.
  18. 18. Chaddad A, Tanougast C, Dandache A, Al Houseini A, Bouridane A. Improving of colon cancer cells detection based on Haralick’s features on segmented histopathological images. 2011. pp. 87–90.
  19. 19. Chaddad A, Maamoun M, Tanougast C, Dandache A. Hardware Implementation of Active Contour Algorithm for Fast Cancer Cells Detection. Biomedical Engineering Conference (SBEC), 2013 29th Southern. 2013. pp. 129–130. 10.1109/SBEC.2013.73
  20. 20. Kim T-Y, Cho N-H, Jeong G-B, Bengtsson E, Choi H-K. 3D texture analysis in renal cell carcinoma tissue image grading. Comput Math Methods Med. 2014;2014: 536217. pmid:25371701
  21. 21. Levin B, Lieberman DA, McFarland B, Smith RA, Brooks D, Andrews KS, et al. Screening and surveillance for the early detection of colorectal cancer and adenomatous polyps, 2008: a joint guideline from the American Cancer Society, the US Multi-Society Task Force on Colorectal Cancer, and the American College of Radiology. CA Cancer J Clin. 2008;58: 130–160. pmid:18322143
  22. 22. Peyret R, Bouridane A, Al-Maadeed SA, Kunhoth S, Khelifi F. Texture analysis for colorectal tumour biopsies using multispectral imagery. 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). 2015. pp. 7218–7221. 10.1109/EMBC.2015.7320057
  23. 23. Chaddad A, Tanougast C, Dandache A, Bouridane A. Extracted haralick’s texture features and morphological parameters from segmented multispectrale texture bio-images for classification of colon cancer cells. WSEAS Trans Biol Biomed. 2011;8: 39–50.
  24. 24. Hiraoka Y, Sedat JW, Agard DA. The use of a charge-coupled device for quantitative optical microscopy of biological structures. Science. 1987;238: 36–41. pmid:3116667
  25. 25. Miller PJ, Hoyt CC. Multispectral imaging with a liquid crystal tunable filter. 1995. pp. 354–365.
  26. 26. Matuszewski BJ, Murphy MF, Burton DR, Marchant TE, Moore CJ, Histace A, et al. Segmentation of cellular structures in actin tagged fluorescence confocal microscopy images. 2011 18th IEEE International Conference on Image Processing (ICIP). 2011. pp. 3081–3084. 10.1109/ICIP.2011.6116316
  27. 27. Histace A, Meziou L, Matuszewski B, Precioso F, Murphy M, Carreiras F. Statistical region based active contour using a fractional entropy descriptor: Application to nuclei cell segmentation in confocal microscopy images. Ann Br Mach Vis Assoc. 2013;2013: 1–15.
  28. 28. Chaddad A, Tanougast C. Real-time abnormal cell detection using a deformable snake model. Health Technol. 2015; 1–9.
  29. 29. 3D Slicer [Internet]. [cited 20 Oct 2014]. Available:
  30. 30. Ganeshan B, Miles KA, Young RCD, Chatwin CR. In search of biologic correlates for liver texture on portal-phase CT. Acad Radiol. 2007;14: 1058–1068. pmid:17707313
  31. 31. Akay M. Introduction: Wavelet transforms in biomedical engineering. Ann Biomed Eng. 1995;23: 529–530. pmid:7503455
  32. 32. Haralick RM, Shanmugam K, Dinstein I. Textural Features for Image Classification. IEEE Trans Syst Man Cybern. 1973;SMC-3: 610–621.
  33. 33. Clausi DA. An analysis of co-occurrence texture statistics as a function of grey level quantization. Can J Remote Sens. 2002;28.
  34. 34. Guo Y, Hastie T, Tibshirani R. Regularized linear discriminant analysis and its application in microarrays. Biostat Oxf Engl. 2007;8: 86–100.
  35. 35. Aggarwal CC. Data Classification: Algorithms and Applications. CRC Press; 2014.
  36. 36. Rokach L. Data Mining with Decision Trees: Theory and Applications. World Scientific; 2007.
  37. 37. Dasarathy BV. Nearest neighbor (NN) norms: nn pattern classification techniques. IEEE Computer Society Press; 1991.
  38. 38. Breiman L, Friedman J, Stone CJ, Olshen RA. Classification and Regression Trees. 1 edition. New York, N.Y.: Chapman and Hall/CRC; 1984.
  39. 39. Witten IH, Frank E. Data Mining: Practical Machine Learning Tools and Techniques, Second Edition. Morgan Kaufmann; 2005.
  40. 40. Refaeilzadeh P, Tang L, Liu H. Cross-Validation. In: LIU L, ÖZSU MT, editors. Encyclopedia of Database Systems. Springer US; 2009. pp. 532–538. Available:
  41. 41. Nichols TE, Holmes AP. Nonparametric permutation tests for functional neuroimaging: a primer with examples. Hum Brain Mapp. 2002;15: 1–25. pmid:11747097
  42. 42. Esgiar AN, Naguib RN, Bennett MK, Murray A. Automated feature extraction and identification of colon carcinoma. Anal Quant Cytol Histol Int Acad Cytol Am Soc Cytol. 1998;20: 297–301.
  43. 43. Jiao L, Chen Q, Li S, Xu Y. Colon Cancer Detection Using Whole Slide Histopathological Images. World Congress on Medical Physics and Biomedical Engineering May 26–31, 2012, Beijing, China. Springer; 2013. pp. 1283–1286. Available:
  44. 44. Rathore S, Hussain M, Ali A, Khan A. A Recent Survey on Colon Cancer Detection Techniques. IEEEACM Trans Comput Biol Bioinforma IEEE ACM. 2013; 4D6431A2-781B-45F6-BF83-00D5F4F52CDD
  45. 45. Xu Y, Mo T, Feng Q, Zhong P, Lai M, Chang EI-C. Deep learning of feature representation with multiple instance learning for medical image analysis. 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2014. pp. 1626–1630. 10.1109/ICASSP.2014.6853873