Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Different Approaches for Extracting Information from the Co-Occurrence Matrix

  • Loris Nanni,

    Affiliation Department of Information Engineering, University of Padua, Padova, Italy

  • Sheryl Brahnam ,

    Affiliation Department of Computer Information Systems, Missouri State University, Springfield, Missouri, United States of America

  • Stefano Ghidoni,

    Affiliation Department of Information Engineering, University of Padua, Padova, Italy

  • Emanuele Menegatti,

    Affiliation Department of Information Engineering, University of Padua, Padova, Italy

  • Tonya Barrier

    Affiliation Department of Computer Information Systems, Missouri State University, Springfield, Missouri, United States of America

Different Approaches for Extracting Information from the Co-Occurrence Matrix

  • Loris Nanni, 
  • Sheryl Brahnam, 
  • Stefano Ghidoni, 
  • Emanuele Menegatti, 
  • Tonya Barrier


In 1979 Haralick famously introduced a method for analyzing the texture of an image: a set of statistics extracted from the co-occurrence matrix. In this paper we investigate novel sets of texture descriptors extracted from the co-occurrence matrix; in addition, we compare and combine different strategies for extending these descriptors. The following approaches are compared: the standard approach proposed by Haralick, two methods that consider the co-occurrence matrix as a three-dimensional shape, a gray-level run-length set of features and the direct use of the co-occurrence matrix projected onto a lower dimensional subspace by principal component analysis. Texture descriptors are extracted from the co-occurrence matrix evaluated at multiple scales. Moreover, the descriptors are extracted not only from the entire co-occurrence matrix but also from subwindows. The resulting texture descriptors are used to train a support vector machine and ensembles. Results show that our novel extraction methods improve the performance of standard methods. We validate our approach across six medical datasets representing different image classification problems using the Wilcoxon signed rank test. The source code used for the approaches tested in this paper will be available at:


When it comes to data, we live in unprecedented times. Across industries and academia, companies and institutions are struggling to handle the massive amounts of data that are accumulating daily. This explosion of data is coming from multiple sources: sensors, social media, the Internet, smart devices, and smart phones. Data stockpiling is motivated in large part by the availability of inexpensive data storage and commodity computer systems that have nearly the computational power of earlier supercomputers [1]. These advances in computing and the economics of ownership have greatly accelerated research in technologies that are producing even more raw data. This is especially the case with computer vision [2] [3] [4], which is now a major component in many applications, ranging from video surveillance software to robotic systems and automatic visual inspection systems for checking industrial products at the end of the production line.

Medical imaging, in particular, is one field that is witnessing rapid technological growth along with a concomitant avalanche of data. To handle this data, specialized research databases (e.g., HUGO, Rfam, and Cancer Cell Map) and metadatabases (e.g., Biograph, mGen, PathogenPortal, and ConsensusPathDB) have been established. Machine vision technology applied to many of these databases has the potential of revolutionizing scientific knowledge in medicine. Already some invaluable gains have been made in the detection of tumors and cancers. For example, in [5] image texture information is utilized to automatically discriminate polyps in colonoscopy images, and in [4] linear discriminant analysis of wavelet features, which has been successfully employed in many nonmedical applications (for instance, traffic accident detection [6] and face identification and verification [7]), is proving highly effective in the detection of tumors in endoscopic images.

Texture analysis is often involved in image classification, but there is no universally recognized definition of texture. It can be viewed as a global pattern arising from the repetition of local subpatterns [8] or as a region where a set of local properties or statistics are either constant, slowly varying, or approximately periodic [9] (for an interesting early catalogue of definitions see [10]). Many different methods for managing texture have been developed that are based on the various ways texture can be characterized. Some of the highest performing methods reported in the literature include the scale-invariant feature transform (SIFT) [11], speeded up robust feature (SURF) [12], histogram of oriented gradients (HOG) [13], gradient location and orientation histogram (GLOH) [14], region covariance matrix (RCM) [15], edgelet [16], gray level co-occurrence matrix (GLCM) [17], local binary patterns (LBP) [18], nonbinary encodings [19], color correlogram (CCG) [20], color coherence vectors (CCV) [21], color indexing [22], steerable filters [23] and Gabor filters [24].

Arguably the Local Binary Pattern (LBP) operator [25] is one of the most powerful approaches utilizing the texture information in an image. LBP is simple, effective, and robust and is proving to be a powerful discriminator in many medical image classification problems. In [26], for example, LBP assigns a Marsh-like score to endoscopical images of pediatric celiac diseases. In [27] a Support Vector Machine (SVM) is coupled with the LBP operator to distinguish real masses from normal parenchyma in mammographic images. LBP has also been combined with other descriptors useful for medical data mining purposes. In [28], for instance, LBP is used to explore brain magnetic resonance data, and in [29] the authors demonstrate how a combination of LBP with other texture descriptors is effective in classifying different cell phenotypes.

In this paper, we focus on improving one of the earliest methods for analyzing texture: the GLCM, originally proposed by Haralick [30] in 1979 for analyzing satellite images. Based on a set of features, or descriptors, that are evaluated starting from a histogram, GLCM is one of the most studied and extensively used general approaches for texture analysis and has recently become the focus of several research groups whose aim is to increase the discriminability of GLCM descriptors. Some interesting recent work in this area includes [31], where the authors consider different values of the distance parameter that influences the GLCM. In [32] features are extracted by weighted summation of GLCM elements from areas presenting high discrimination. In [33] descriptors are obtained from the gray level-gradient co-occurrence matrix, formed by calculating the gradient value of each pixel in the neighborhood of interest points. In [34] GLCM is combined with the edge orientation co-occurrence matrix of superior order [35], thereby taking into consideration both the gray levels of the image pixels and such local features as edges. Multi-scale analysis has also been performed using the GLCM. For example, in [36] and in [37] multiple scales are considered by changing the window size used to extract the GLCM descriptors, and in [38] the image is rescaled to different sizes and co-occurrence descriptors are extracted from each rescaling.

Recently the addition of color information has been evaluated for co-occurrence matrices [39]. In [40], for instance, a colors-gradient co-occurrence matrix (CGCM) is proposed from which 27-dimensional statistical features are extracted. In [41] superior results are obtained on content-based image retrieval by employing a combination of the contourlet transform and the GLCM. First, the contourlet transform is performed for four subbands of the image, and then GLCM features are extracted from each band. In [42] the authors propose the Gradient Magnitude based Angle Co- occurrence Matrix for color image classification that is based on three different types of gradients defined in the RGB space.

Finally, some very recent work has been proposed that combines LBP with GLCM. For example, in [43], GLCM is constructed after LBP is applied to the image. Features are then extracted from second-order statistical parameters of the resulting LBGLCM matrix. Similarly, in [44], an LBP image is built after performing a Gaussian filtering pyramid preprocessing step, and the GLCM is constructed from the resulting LBP image.

A major difficulty encountered when analyzing texture is that results strongly depend on image resolution and scale, an effect that is especially problematic with edge-based approaches. One of the goals of this work is to assess the performance improvement that can be gained using a multi-scale approach. In the literature, several authors have recently reported combining several multi-scale approaches with LBP descriptors (see, for instance, [45] and [46]). We show in this study that it is also possible to improve the performance of different descriptors extracted from the co-occurrence matrix by coupling them with a multi-scale approach.

Our main intention, however, is to compare different methods for extracting features given the co-occurrence matrix (we are not comparing and combining different pre-processing methods applied before the co-occurrence matrix extraction). As described in detail in section 2, the following extraction methods are investigated:

  • Extracting descriptors using the standard approach proposed by Haralick [17];
  • Extracting a set of 3D descriptors by considering the co-occurrence matrix as a 3D shape [47];
  • Extracting descriptors from different 2D shapes by considering the co-occurrence matrix as a 3D shape: The 2D shapes are obtained by intersecting the co-occurrence matrix with a set of horizontal planes at given heights;
  • Extracting gray-level run-length features [48];
  • Directly using the co-occurrence matrix as a descriptor by projecting it onto a lower dimensional subspace using principal component analysis;

We improve the performance of these descriptors further by extracting them not only from the entire original co-occurrence matrix, but also from multiple scales of the original image obtained by Gaussian filtering and by dividing the original co-occurrence matrix into different subwindows and extracting features separately from each subwindow. These descriptors are then used to train separate SVM classifiers whose results are combined by sum rule or weighted sum rule.

The experimental results presented in section 4 show that our new approach improves the performance of standard methods and some state-of-the-art approaches. We validate our approach across six medical datasets, described in section 3, representing very different image classification problems.

Proposed System

The focus of this paper is on extracting descriptors from the co-occurrence matrix with the goal of enhancing the performance of Haralick’s descriptors. Improvements are achieved by applying the multi-scale approach that overcomes the main weakness of texture-based features (i.e., the dependency on scale discussed in the introduction) and by extracting features from different subwindows of the co-occurrence matrix and then combining them.

Each of the approaches tested in the experimental section is explained below. In all experiments, SVM [49][51] is used as the base classifier. In this study, linear, polynomial and radial basis function kernels are tested. For each dataset, the best kernel and the best set of parameters are chosen using a 5-fold cross validation approach on the training data.

The Multi-scale Approach

Using this approach, images are generated by applying a 2D symmetric Gaussian lowpass filter of size k (we use k = 3 and k = 5) with standard deviation 1. As illustrated in Figure 1, the original image is filtered to obtain a set of smoothed versions of the original image.

Figure 1. Images illustrating the effect of the multi-scale approach.

The GLDM Co-occurrence Matrix

GLDM [17], [30] is a particular type of co-occurrence matrix obtained as the histogram on a 2D domain of dimension NGL×NGL, where NGL is the number of gray levels in the image (typically 256). In other words, the co-occurrence matrix counts the number of gray level transitions between two pixel values such that the bin of the histogram whose coordinates are equal to the values of the two pixels is incremented. The way pixel couples are determined depends on the two parameters, d and θ. A value of d = 1 and θ = 0, for example, would produce pixel couples that are adjacent to each other on the same row. In our experiments four directions are considered: the horizontal (H), the vertical (V), the diagonal top left-bottom right, or right-down (RD), and the top right-bottom left, or left-down (LD).

The Standard Haralick Statistics

The idea of using statistical indicators was originally proposed in [30]. In the experimental section, we label this approach HAR. The following HAR indicators are evaluated:

  1. Energy
  2. Correlation
  3. Inertia
  4. Entropy
  5. Inverse difference moment
  6. Sum average
  7. Sum variance
  8. Sum entropy
  9. Difference average
  10. Difference variance
  11. Difference entropy
  12. Information measure of correlation 1
  13. Information measure of correlation 2.

A set of 13 descriptors is calculated from each co-occurrence matrix evaluated at θ = {0°, 45°, 90°, 135°} and with distance d = {1, 3}. A descriptor is obtained by concatenating the features extracted for each distance and orientation value.

In our experimental section, we also report the performance of several HAR variants:

  • HR: features are extracted from the whole image only (note: several comparisons of different parameters settings for HR are reported in [47], but in this paper we use only the best configuration reported in that study).
  • HRsub: a feature set is extracted from the whole co-occurrence matrix and from each subwindow (we use four subwindows in this work). Each set of features is then used to train a separate SVM. All 5 SVMs are combined by weighted sum rule, with weight of 4 for the SVM trained on the whole matrix, and weight of 1 for the others fours SVMs. To avoid presenting huge table, only the results of the best four subwindows are presented, defined using the coordinates (0, 0) to (127, 127); (128, 128) to (255, 255); (0, 128) to (255, 128); and (0, 128) to (128, 255);
  • HRsca, features are extracted from the original image and the two filtered images The 15 SVMs are combined by weighted sum rule, with weight of the SVMs trained using features extracted from the original image given the value of 4, while the other SVMs have a weight of 1.


The SHAPE approach explores the shape of the co-occurrence matrix by considering it as a 3D function (see Figure 2). This approach has been explored in detail in [52], [47], and [53]. The main idea of SHAPE is to intersect the GLDM with a set of horizontal planes at given heights and then derive a set of features based on the contours of the intersection, which defines a complex shape made up of one or more extractable blobs. The blob with the largest area, referred to as the main blob, is selected for extracting features and is fitted to an ellipse in order to simplify analysis. Although this approximation of the main blob shape to an ellipse results in some information loss, it offers the advantage of making the comparison among curves much easier.

Figure 2. Illustration of the co-occurrence matrix as a 3D function.

Level curves are considered towards the base of the co-occurrence matrix, starting at height 1 and then going until height 19, with a distance of 2 between two consecutive planes. Level curves are all at a relatively low height because that region is very stable unlike the upper part of the co-occurrence matrix, which is much more unstable because of image noise. For this reason the co-occurrence matrix is not normalized, since normalization to the highest bin would introduce instabilities. Other types of normalization could be performed with respect to the total volume of the co-occurrence matrix, but this depends on the size of the original image, which is constant in most cases, making the normalization useless.

For each level, a set of descriptors extracted from the ellipses derived from the co-occurrence matrix is evaluated. The features describing all levels are then jointly analyzed for extracting a set of nine features that describe the evolution of the level curves (see [47], [52] for details).

These features are used to provide a characterization of the input image and can be directly used as input for a classifier. This is the principle exploited in the HAR approach. In the case of the SHAPE approach, however, features are evaluated not only on the entire co-occurrence matrix (as in [52]) but also on 12 subwindows of the GLDM defined by the following coordinates: #1: (0, 0) to (127, 127); #2: (128, 128) to (255, 255); #3: (0, 0) to (191, 191); #4: (64, 64) to (255, 255); #5: (0, 0) to (95, 95); #6: (31, 31) to (95, 95); #7: (63, 63) to (127, 127); #8: (95, 95) to (159, 159); #9: (127, 127) to (191, 191); #10: (159, 159) to (223, 223); #11: (191, 191) to (255, 255); and #12: (63, 63) to (191, 191). Several experiments using the entire GLDM along with these same subwindows are reported in [47]. (Note: in the case of SHAPE, we use more than the four subwindows because the performance of SHAPE improves when smaller subwindows are used; this is not the case with the other methods explored in this work).

For each of these 13 windows (counting the GLMD as a whole along with the 12 subwindows) a different feature vector is extracted, and these 13 descriptors are used to train 13 separate SVMs. Results are combined by weighted sum rule, where a weight of 1 is assigned to the first five descriptors, and a weight of 0.5 is assigned to the remainder. Each set of 13 descriptors is derived from co-occurrence matrices evaluated at θ = {0°, 45°, 90°, 135°} and, in the case when multiple values are used for the distance, d = {1, 3}. The feature vector is obtained by concatenating the features extracted for each value of the distance. In the experimental section, SH refers to the case where features are extracted from the entire co-occurrence matrix only, while SHsub is the method based on all 13 windows. When SHsub is coupled with the multi-scale approach, we call this combination SHsca.

Curvature (CR)

The curvature (CR) algorithm [54] takes as input a set of edge point samples and a set of circular masks with varying radii. CR counts the number of samples falling in each mask when the mask is centered on edge points representing a contour. This count is used to obtain a measure of curvature, and these curvature measures are then quantized in a histogram of bins of equal size which are collected to form a feature vector.

The CR algorithm is applied to the co-occurrence matrix considering a set of level curves starting at height 1, going until height 15, with a distance of 3 between two consecutive planes. For each 2D shape defined for a given height, the descriptors are extracted from co-occurrence matrices evaluated at θ = {0°, 45°, 90°, 135°}, and multiple values of d = {1, 3}. The descriptor is obtained by concatenating the features extracted for each value of the distance and of the orientation. In the experimental section, we also report the performance of the following GL variants:

  • CU: as in HR but using the CR features;
  • CUsub: as in HRsub but using the CR features;
  • CUsca: as in HRsca but using the CR features.

1.5 Gray-level Run-length Features (GL)

GL [48] derives descriptors from a run-length matrix that is based on characteristics of the gray level runs within a given image. A gray level run is a set of consecutive pixels with the same value, and the run length is the number of pixels in the set. The run-length matrix P contains in each location p(i, j) the number of runs of length j at a given gray level i. Starting from the run-length matrix it is possible to obtain several indicators, as explained in [48]. In our experiments, we consider the following:

  • Short Run Emphasis (SRE), evaluated as:

where is the total number of runs, the number of gray levels, and the maximum run length.

  • Long Run Emphasis (LRE):

  • Gray Level Nonuniformity (GLN):

  • Run Length Nonniformity (RLN):

  • Run Percentage (RP):

where is the total number of pixels in the image.

  • Low Grey-Level Run Emphasis (LGRE):

  • High Grey-Level Run Emphasis (HGRE):

  • Short Run Low Grey Level Emphasis (SRLGE):

  • Short Run High Grey Level Emphasis (SRHGE):

  • Long Run Low Grey Level Emphasis (LRLGE):

  • Long Run High Grey Level Emphasis (LRHGE):

In our system we calculate the indicators described above from a run-length matrix that is evaluated on the GLDM. Since multiple GLDMs are calculated at several values of θ = {0°, 45°, 90°, 135°} and several values of d = {1, 3}, the GL descriptor is obtained by concatenating all the described features for all values of θ and d. The GL approach in turn has its own orientation: all values considered in [48] are evaluated in our system, namely θGL = {0°, 45°, 90°, 135°}.

In the experimental section, we also report the performance of the following GL variants:

  • GR: as in HR but using the GL features;
  • GRsub: as in HRsub but using the GL features;
  • GRsca: as in HRsca but using the GL features.

Lower Dimensional Subspace (LDS)

In our experiments, we also use the co-occurrence matrix itself as a descriptor. Since it is very high in dimensionality, we project it onto a lower dimensional subspace using PCA, where 99% of the variance is retained for input into an SVM. As in the previous approaches, each set of features is extracted from co-occurrence matrices evaluated at θ = {0°, 45°, 90°, 135°}, and multiple values are used for the distance, d = {1, 3}. The descriptor is obtained by concatenating the features extracted for each value of the distance and of the orientation.

The use the following approaches with LDS:

  1. LD: as in HR but using the LDS features;
  2. LDsub: as in HRsub but using the LDS features;
  3. LDsca: as in HRsca but using the LDS features.


We validate our systems across six medical datasets to assess generality. These datasets represent different computer vision problems. As an example of image diversity, Figure 3 shows the different visual characteristics of sample images representative of our six datasets.

Figure 3. Illustration of image diversity with a sample image representative of each of the six datasets.

The datasets used for evaluating approaches are the following (notice that the RGB images are converted in gray level images):

  • PAP: this Pap Smear dataset [55] contains images representing cells that are used in the diagnosis of cervical cancer.
  • VIR: this dataset [56] contains images of viruses extracted using negative stain transmission electron microscopy. We use the 10-fold validation division of images shared by the authors. However, we do not exploit their mask for background subtraction. Instead, we use the entire image for extracting features as this method produced better results.
  • HI: this Histopatology dataset [57] is composed of images from different organs that are representative of the four fundamental tissues.
  • BC: this dataset [58] contains 273 malignant and 311 benign breast cancer images.
  • PR: this dataset, developed by [59], contains 118 DNA-binding Proteins and 231 Non-DNA-binding proteins. Texture descriptors are extracted from the 2D distance matrix that represents each protein, and this is obtained from the 3D tertiary structure of a given protein (considering only atoms that belong to the protein backbone, see [59] for details).
  • CHO: this cell dataset [60] that contains 327 fluorescent microscopy images taken from Chinese Hamster Ovary cells and belonging to five different classes. Images are 16 bit grayscale of size 512×382 pixels.

A descriptive summary of each dataset along with website links to each dataset is reported in Table 1.


The 5-fold cross-validation protocol is used for testing each texture descriptor, with the exception of the VIR dataset, for which the original testing protocol is used. The area under the ROC curve (AUC) is used as the performance indicator because it provides a better overview of classification results. AUC is a scalar measure that can be interpreted as the probability that the classifier will assign a higher score to a randomly picked positive sample than to a randomly picked negative sample [61]. In the multi-class problem, AUC is calculated using the one-versus-all approach (a given class is considered as “positive” and all the other classes are considered as “negative”) and the average AUC is reported.

The aim of the first experiment, see table 2, is to establish the usefulness of extracting features not only from the whole co-occurrence matrix but also from different sub-windows.

Table 2. Table 2. Usefulness of extracting features from the co-occurrence matrix subwindows.

The aim of the second experiment, see table 3, is to show the performance gains that can be achieved using the multi-scale approach.

Table 3. Table 3. Performance gains using the multi-scale approach.

Examining Tables 2 and 3, the following conclusions can be drawn:

  • It is clear that all methods improve when features are also extracted from GLCM subwindows; even the standard HR improves when coupled with subwindow extraction (to the best of our knowledge, this is the first paper to explore using subwindows to extract information from the co-occurrence matrix);
  • Coupling approaches with the multi-scale approach further improves results;
  • The novel method GRsca obtains a performance similar to HRsca;
  • The ensemble proposed here, HRsca, outperforms the base HR;
  • CUsca and LDsca work well on some datasets and not on others: when using these approaches, it would be desirable to test them using the training data to determine whether they are suited to a specific task. For instance, CU is more suited for binary images than for texture images. LD is not well suited for extracting the local information found in the co-occurrence matrix; nonetheless, LD is able to extract some useful features.

To statistically validate of our experiments, we used the Wilcoxon signed rank test [62]:

  • HRsca outperforms HR (p-value 0.05);
  • GRsca outperforms GR (p-value 0.05);
  • HRsca and GRsca obtain a similar performance.

Before turning to the next set of experiments, it should be observed that the results of CU and LD highlight the difficulty of predicting which descriptors will work well for a given dataset; the best way to determine this is to use the training set of each dataset to assess performance. It is well-known in the computer vision and machine learning communities that no stand-alone approach is consistently superior to any other. The “no free lunch” theorem states that any two learning algorithms exhibit the same performance when error rates are averaged across all possible problem sets (see, e.g., [63]). In our opinion, the main value of combining different descriptors is that they offer the most feasible way of coping with the “no free lunch” theorem.

The aim of the third experiment, see table 4, is to show the performance gain that is possible by fusing different descriptors extracted from the co-occurrence matrix. The descriptors chosen for this experiment were the following:

  • Old: the best ensemble proposed in [52] based on features extracted from co-occurrence matrix;
  • SUM2: sum rule between HRsca and GRsca. Notice: when we combine the SVMs trained with different descriptors, the scores of each SVM are always normalized to mean 0 and standard deviation 1, before the fusion step.
  • WS2: weighted sum rule between HRsca (weight 2) and GRsca (weight 1);
  • W = 2: weighted sum rule between SUM2 (weight 2) and SHsca (weight 1);
  • W = 3; weighted sum rule between SUM2 (weight 3) and SHsca (weight 1);

The best result was obtained by choosing W = 3, which confirms that all three methods (HRsca, GRsca and SHsca) extract different information from the co-occurrence matrix. W = 3 outperforms HRsca with p-value 0.05 using Wilcoxon signed rank test.

For comparison purposes, in Table 5, we report the results of some of the best performing texture descriptors reported in the literature:

  • Local binary patterns (LBP) [25];
  • Local ternary patterns (LTP) [64];
  • Multi-threshold local quinary coding (MT) [53].

In Table 5 we also report the following fusion method:

  • W1, sum rule between MT and W = 3.

It is interesting to note that the performance of the HRsca method that we propose is similar to what can be achieved using MT. Moreover, W1 (the ensemble of different descriptors) outperforms all the base approaches that belong to it, leading to a very high performing set of descriptors.

Finally, we perform experiments using SPIN features [64] considering the GLDM as a 3D image. This method extracts reliable features, but the performance is lower than other methods reported in this work, e.g., HRsub. Using SPIN features considering the GLDM as a 3D image obtains the following results: PAP = 77.2; BC = 71.8; PR = 84.2; CHO = 93.0).


In this study we extended previous work reported in the literature on texture analysis techniques based on the co-occurrence matrix. We compared and combined different strategies for extending the texture descriptors extracted from the co-occurrence matrix. These methods were further improved by combining them with a multi-scale approach based on Gaussian filtering and by extracting features not only from the entire co-occurrence matrix but also from subwindows. Moreover, we proposed a novel set of features (Grey-level run-length features) to use with the co-occurrence matrix. We also showed that our ensemble approach improves the performance of SHAPE and standard Haralick-based features and outperforms other stand-alone approaches without ad hoc dataset parameter tuning. Our proposed system was validated across six different medical image classification problems, thereby demonstrating the generality of our approach. Our results were also compared with some state-of-the-art descriptors. For all experiments SVM was used as the base classifier.

In our opinion the most valuable result of this paper is demonstrating that it is possible to extract even more information from the co-occurrence matrix than has been extracted thus far. Our study shows that it is worthwhile exploring more techniques for deriving new descriptors from the co-occurrence matrix. Future experiments, for instance, might examine different methods for processing the image before extracting the co-occurrence matrix. We hypothesize that Gabor filters and wavelet decomposition will prove to be valuable preprocessing methods.

Author Contributions

Conceived and designed the experiments: LN SB SG EM TB. Performed the experiments: LN. Analyzed the data: LN SB SG EM TB. Contributed reagents/materials/analysis tools: LN SB SG EM TB. Wrote the paper: LN SB SG EM TB.


  1. 1. Zikopoulos PC, Eaton C, deRoos D, Deutsch T, Lapis G (2012) Understanding big data: Analytics for enterprise class hadoop and streaming data. New York: McGraw-Hill.
  2. 2. Hamilton NA, Wang JTH, Kerr MC, Teasdale RD (2009) Statistical and visual differentiation of subcellular imaging. BMC Bioinformatics 10.
  3. 3. Murphy RF (2006) Putting proteins on the map. Nature Biotechnology 24: 1223–1224.
  4. 4. Karkanis SA, Iakovidis DK, Maroulis DE, Karras DA, Tzivras M (2003) Computer-aided tumor detection in endoscopic video using color wavelet features. IEEE Transaction on Information Technology in Biomedicine 7: 141–152.
  5. 5. Ameling S, Wirth S, Paulus D, Lacey G, Vilarino F (2009) Texture-based polyp detection in colonoscopy. In: Meinzer H-P, Deserno TM, Handels H, Tolxdorff T, editors. Bildverarbeitung für die Medizin. Berlin and Heidelberg, Germany: Springer. 346–350.
  6. 6. Samant A, Adeli H (2000) Feature extraction for traffic incident detection using wavelet transform and linear discriminant analysis. Computer-Aided Civil and Infrastructure Engineering 15: 241–250.
  7. 7. Shen L, Bai L, Fairhurst M (2007) Gabor wavelets and general discriminant analysis for face identification and verification. Image and Vision Computing 25: 553–563.
  8. 8. Zucker SW (1976) Towards a model of texture. Computer Graphics Image Processiong 5: 190–202.
  9. 9. Sklansky J (1978) Image segmentation and feature extraction. IEEE Transactions on Systems, Man, and Cybernetics SMC-8: 237–247.
  10. 10. Coggins JM (1982) A framework for texture analysis based on spatial filtering [Ph.D. Thesis]. East Lansing, Michigan: Michigan State University.
  11. 11. Lowe DG (2004) Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60: 91–110.
  12. 12. Bay H, Tuytelaars T, Gool LV (2006) SURF: Speeded up robust features. European Conference on Computer Vision 1: 404–417.
  13. 13. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. 9th European Conference on Computer Vision. San Diego, CA.
  14. 14. Mikolajczyk K, Schmid C (2005) A performance evaluation of local descriptors. IEEE Transactions on Pattern Analysis and Machine Intelligence 29: 1615–1630.
  15. 15. Tuzel O, Porikli F, Meer P (2006) Region covariance: A fast descriptor for detection and classi?Cation. 9th European Conference on Computer Vision. 589–600.
  16. 16. Wu B, Nevatia R (2005) Detection of multiple, partially occluded humans in a single image by bayesian combination of edgelet part detectors. IEEE International Conference on Computer Vision. 90–97.
  17. 17. Haralick RM, Shanmugam K, Dinstein I (1973) Textural features for image classification. IEEE Transactions on Systems, Man, and Cybernetics 3: 610–621.
  18. 18. Ojala T, Pietikäinen M, Harwood D (1996) A comparative study of texture measures with classification based on featured distribution. Pattern Recognition Letters 29: 51–59.
  19. 19. Paci M, Nanni L, Lathi A, Aalto-Setälä K, Hyttinen J, et al. (2013) Non-binary coding for texture descriptors in sub-cellular and stem cell image classification. Current Bioinformatics 8: 208–219.
  20. 20. Huang J, Kumar SR, Mitra M, Zhu WJ, Zabih R (1997) Image indexing using color correlograms. IEEE Conference on Computer Vision and Pattern Recognition. San Jose, Puerto Rico. 762–768.
  21. 21. Pass G, Zabih R, Miller J (1996) Comparing images using color coherence vectors. ACM international conference on Multimedia. 65–73.
  22. 22. Swain MJ, Ballard DH (1991) Color indexing. International Journal of Computer Vision 7: 11–32.
  23. 23. Freeman WT, Adelson EH (1991) The design and use of steerable filters. IEEE Transactions on Pattern Analysis and Machine Intelligence 13: 891–906.
  24. 24. Jain AK, Farrokhnia F (1991) Unsupervised texture segmentation using gabor filters. Pattern Recognition Letters 24: 1167–1186.
  25. 25. Ojala T, Pietikainen M, Maeenpaa T (2002) Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence 24: 971–987.
  26. 26. Vécsei A, Amann G, Hegenbart S, Liedlgruber M, Uhl A (2011) Automated marsh-like classification of celiac disease in children using local texture operators. Computers in Biology and Medicine 41: 313–325.
  27. 27. Oliver A, Lladó X, Freixenet J, Martí J (2007) False positive reduction in mammographic mass detection using local binary patterns. Medical Image Computing and Computer-Assisted Intervention (MICCAI), Lecture Notes in Computer Science 4791. Brisbane, Australia: Springer. 286–293.
  28. 28. Unay D, Ekin A (2008) Intensity versus texture for medical image search and retrieval 5th IEEE International Symposium on Biomedical Imaging: From Nano to Macro. 241–244.
  29. 29. Nanni L, Lumini A (2008) A reliable method for cell phenotype image classification. Artificial Intelligence in Medicine 43: 87–97.
  30. 30. Haralick RM (1979) Statistical and structural approaches to texture. Proceedings of the IEEE 67: 786–804.
  31. 31. Gelzinis A, Verikas A, Bacauskiene M (2007) Increasing the discrimination power of the co-occurrence matrix-based features. Pattern Recognition 40: 2367–2372.
  32. 32. Walker R, Jackway P, Longstaff D (2003) Genetic algorithm optimization of adaptive multi-scale GLCM features. International Journal of Pattern Recognition and Artificial Intelligence 17: 17–39.
  33. 33. Chen S, Chengdong W, Chen D, Tan W (2009) Scene classification based on gray level-gradient co-occurrence matrix in the neighborhood of interest points. IEEE International Conference on Intelligent Computing and Intelligent Systems (ICIS 2009) 482–485.
  34. 34. Mitrea D, Mitrea P, Nedevschi S, Badea R, Lupsor M, et al.. (2012) Abdominal tumor characterization and recognition using superior-order cooccurrence matrices, based on ultrasound images. Computational and Mathematical Methods in Medicine.
  35. 35. Akono A, Tankam NT, Tonye E (2005) Nouvel algorithme d’evaluation de texture d’ordre n sur la classification de l’ocupation des sols de la region volcanique du mont Cameroun. Teledetection 5: 227–244.
  36. 36. Hu Y (2009) Unsupervised texture classification by combining multi-scale features and k-means classifier. Chinese Conference on Pattern Recognition. 1–5.
  37. 37. Pacifici F, Chini M (2008) Urban land-use multi-scale textural analysis. IEEE International Geoscience and Remote Sensing Symposium. 342–345.
  38. 38. Rakwatin P, Longepe N, Isoguchi O, Shimada M, Uryu Y (2010) Mapping tropical forest using ALOS PALSAR 50 m resolution data with multiscale GLCM analysis. IEEE International Geoscience and Remote Sensing Symposium. 1234–1237.
  39. 39. Benco M, Hudec R (2007) Novel method for color textures features extraction based on GLCM. Radioengineering 4: 64–67.
  40. 40. Gong R, Wang H (2012) Steganalysis for GIF images based on colors-gradient co-occurrence matrix. Optics Communications 285: 4961–4965.
  41. 41. Nguyen-Duc H, Do-Hong T, Le-Tien T, Bui-Thu C (2010) A new descriptor for image retrieval using contourlet co-occurrence. Third International Conference on Communications and Electronics (ICCE). 169–174.
  42. 42. Zhang R, Yin B, Zhao Q, Yang B (2010) An efficient color image classification method using gradient magnitude based angle cooccurrence matrix. IEEE 17th International Conference on Image Processing. Hong Kong. 1073–1076.
  43. 43. Sastry SS, Kumari TV, Rao CN, Mallika K, Lakshminarayana S, et al. (2012) Transition temperatures of thermotropic liquid crystals fromthe local binary gray level co-occurrence matrix. Advances in Condensed Matter Physics 2012: 1–9.
  44. 44. Sun X, Wang J, Chen R, She MFH, Kong L (2012) Multi-scale local pattern co-occurrence matrix for textural image classification. IEEE World Congress on Computational Intelligence. Brisbane, Australia.
  45. 45. Ren X-M, Wang X-F, Zhao Y (2012) An efficient multi-scale overlapped block LBP approach for leaf image recognition. Intelligent Computing Theories and Applications. 237–243.
  46. 46. Qian X, Hua X-S, Chen P, Ke L (2011) PLBP: An effective local binary patterns texture descriptor with pyramid representation. Pattern Recognition Letters 44: 2502–2515.
  47. 47. Nanni L, Ghidoni S, Menegatti E (2013) A comparison of multi-scale approaches for extracting image descriptors from the co-occurrence matrix. Computer Communication & Collaboration. In press.
  48. 48. Tang X (1998) Texture information in run-length matrices. IEEE Transactions On Image Processing. 1602–1609.
  49. 49. Vapnik V, Lerner A (1963) Pattern recognition using generalized portrait method. Automation and Remote Control 24: 774–780.
  50. 50. Vapnik V, Chervonenkis A (1964) A note on one class of perceptrons. Automation and Remote Control 25.
  51. 51. Kuncheva LI, Whitaker CJ (2003) Measures of Diversity in Classifier Ensembles and their Relationship with the ensemble accuracy. Machine Learning 51: 181–207.
  52. 52. Ghidoni S, Cielniak G, Menegatti E (2012) Texture-based crowd detection and localisation. International Conference on Intelligent Autonomous Systems (IAS-12). Jeju Island, Korea.
  53. 53. Nanni L, Paci M, Brahnam S, Ghidoni S, Menegatti E (2013) Local phase quantization texture descriptor for protein classification. The 14th International Conference on Bioinformatics and Computational Biology (BIOCOMP’13). Las Vegas, NV. 56–61.
  54. 54. Dominio F, Donadeo M, Zanuttigh P (2013) Depth gives a hand to gesture recognition. Pattern Recognition Letters. In press.
  55. 55. Jantzen J, Norup J, Dounias G, Bjerregaard B (2005) Pap-smear benchmark data for pattern classification. Nature inspired Smart Information Systems (NiSIS). Albufeira, Portugal. 1–9.
  56. 56. Kylberg G, Uppström M, Sintorn I-M (2011) Virus texture analysis using local binary patterns and radial density profiles. In: Martin S, Kim S-W, editors. 18th Iberoamerican Congress on Pattern Recognition (CIARP). 573–580.
  57. 57. Cruz-Roa A, Caicedo JC, González FA (2011) Visual pattern mining in histology image collections using bag of features. Artificial Intelligence in Medicine: 91–106.
  58. 58. Junior GB, Cardoso de Paiva A, Silva AC, Muniz de Oliveira AC (2009) Classification of breast tissues using Moran’s index and Geary’s coefficient as texture signatures and SVM. Computers in Biology and Medicine 39: 1063–1072.
  59. 59. Nanni L, Shi J-Y, Brahnam S, Lumini A (2010) Protein classification using texture descriptors extracted from the protein backbone image. Journal of Theoretical Biology 3: 1024–1032.
  60. 60. Boland MV, Markey MK, Murphy RF (1998) Automated recognition of patterns characteristic of subcellular structures in fluorescence microscopy images. Cytopathology 33: 366–375.
  61. 61. Fawcett T (2004) ROC graphs: Notes and practical considerations for researchers. Palo Alto, USA: HP Laboratories.
  62. 62. Demšar J (2006) Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research 7 1–30.
  63. 63. Wolpert DH (2001) The supervised learning no-free-lunch theorems. 6th Online World Conference on Soft Computing in Industrial Applications. 25–42.
  64. 64. Tan X, Triggs B (2007) Enhanced local texture feature sets for face recognition under difficult lighting conditions. Analysis and Modelling of Faces and Gestures LNCS 4778: 168–182.