Figures
Abstract
This study introduces an innovative approach to image classification that uses Gaussian copulas with an Empirical Cumulative Distribution Function (ECDF) approach. The strategic use of distribution functions as feature descriptors simplifies the approach and enables a better understanding of the correlation structure between features in the image. This approach helps the model understand the contextual relationships between different parts of the image, resulting in a more abstract representation than a direct representation of individual pixel values. The proposed model utilizes the Distribution Function of the Distribution Value (DFDV) as the margin distribution. The Modified National Institute of Standards and Technology (MNIST) dataset is comprehensively used to assess the effectiveness of this model. The results show that the model achieves a noteworthy level of accuracy, with an average accuracy of 62.22% and a peak accuracy of 96.92%. This success was obtained by applying the Inference Function for Marginals (IFM) principles during the training stage.
Citation: Indratno SW, Winarni S, Sari KN (2024) Classification of images using Gaussian copula model in empirical cumulative distribution function space. PLoS ONE 19(12): e0309884. https://doi.org/10.1371/journal.pone.0309884
Editor: Sadiq H. Abdulhussain, University of Baghdad, IRAQ
Received: August 18, 2023; Accepted: August 21, 2024; Published: December 6, 2024
Copyright: © 2024 Indratno et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The dataset is freely available at https://www.kaggle.com/datasets/hojjatk/mnist-dataset.
Funding: This research was funded by the Bandung Institute of Technology International Research Program in 2023 titled "Development of Deep Gaussian Process”. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Image classification constitutes a fundamental aspect of computer vision, encompassing the systematic categorization of images into distinct classes or categories based on their discernible visual features and characteristics. The modeling process involves acquiring knowledge through pattern learning from the training data. The program’s efficacy in capturing these data patterns is directly contingent upon the magnitude of the training dataset [1]. The practical applications of image classification are diverse, spanning tasks such as object identification [2], facial detection [3], handwriting recognition [4–8], automated automotive number plate recognition [9], and medical condition diagnosis [10]. Nevertheless, image classification presents a formidable challenge due to the necessity of differentiating between objects and patterns with visual similarities. To address this challenge, meaningful features, including shape, texture, color, or a combination thereof, are judiciously extracted from the images [11]. These extracted features are then employed to train a machine learning model, facilitating the acquisition of the capability to discern patterns and make predictions for new, previously unseen images. Consequently, image classification emerges as a critical component in automating tasks that would otherwise demand human intervention.
Developing a model for image classification is an engaging pursuit, and various methodologies have been formulated to confront this challenge. Prior investigations have scrutinized the identification of image shapes and textures, juxtaposing the Support Vector Machine (SVM) technique against K-Nearest Neighbor (KNN) methods [12]. Subsequent inquiries have expanded into the domain of Support Kernel Machine (SKM), Probabilistic Neural Network (NN) [13, 14], SVM, and Artificial Neural Network (ANN) methodologies [15, 16]. Nevertheless, contemporary strides in image classification techniques have shifted toward deep learning, prominently employing Convolutional Neural Networks (CNN) [17–19]. Despite the encouraging outcomes showcased by CNN, the complexity of the model may introduce practical challenges in its implementation [20–22]. As the complexity of deep learning models increases, interpreting them becomes progressively complex [23–25].
Several studies have examined applying Gaussian copulas as classification models for image analysis. For instance, a study utilized a Gaussian copula model with the Gamma distribution as the marginal distribution to classify image textures in the Vision Texture database [26]. Another investigation introduced a Gaussian copula-based probabilistic classifier to distinguish between object and background images [27]. Additionally, a research endeavor employed the copula distribution function and a diagnostic classifier to categorize brain image data [28]. The application of Gaussian copulas has extended to statistical pattern recognition for acute inflammation data [29] and the classification of hyperspectral images [30]. Several studies incorporated transformations as inputs to the Gaussian copula model [31–33], with all investigations using pixel intensity values as inputs. Although these studies have shown promising results, the reliance on pixel intensity values may pose intensive challenges to extracting more abstract features. Therefore, this study proposes a Gaussian copula with an Empirical Cumulative Distribution Function (ECDF) approach for extracting image features. This study represents an advancement in research focused on image classification using symbolic data [34–36].
This study introduces an effective new approach to image classification by combining the Gaussian copula model and applying the cumulative distribution function strategy. Compared to previous research in image classification using Gaussian copula models, this approach is a step forward that aims to simplify the methodology by exploiting the distribution of pixels in an image to describe its characteristics accurately. This approach helps the model understand the contextual relationships between elements in the image, resulting in a more abstract representation of the image than that obtained from individual pixel values. In addition, by using distributions as features, memory to store distribution parameters is much more efficient than the memory required to directly store many image pixel values.
The research methodology systematically divides each image into p different partitions, where the Empirical Cumulative Distribution Function (ECDF) is computed within each partition, generating a set of ECDF functions for N images. At a specified point Tt the ECDF values from each function are amalgamated to construct a Distribution Function of Distribution Values (DFDV), effectively encapsulating the distinct characteristics of the image object. Selecting s points in one partition yields s points are taken in one partition, this results in s points in one partition yields s functions to comprehensively represent the image. These partitions originate from the same image, so the interdependence is modeled using a Gaussian copula joint distribution function. The primary objective of this research is to establish a classification model grounded in the Gaussian copula, employing a cumulative distribution function space approach. This strategy aims to enhance the clarity and efficiency of image classification by meticulously analyzing distribution functions derived from partitioned image data.
Research methodology
The Gaussian copula model using ECDF and DFDV
Suppose there are N images denoted by X1,X2,…,XN, where Each of these images is partitioned into p equal sized-sections. Let Xj,i represent random variables indicating pixel values in partition j and image i, where j = 1,2,…,p and i = 1,2,…,N. For each partition an empirical cumulative distribution of the pixel values can be derived as follows:
(1)
where Mj,i(x) is the count of pixel values in partition j of image i that are less than or equal to x. Therefore, one gets the following collection of empirical cumulative distribution of the partition j:
(2)
Note that ℱj,i(x)∈[0,1] for all x.
The collection of empirical cumulative distribution allows us to investigate the character of partition j further. Let us consider specific points , where
. From (2) one gets
. Therefore, one can create a mapping from the interval [0,1] to [0,1] as follows:
(3)
Here, ℱj,i(T) represents the value of the ECDF for partition j of image i at point T [36]. Since is a cumulative distribution of empirical cumulative distribution values at
it is called as Distribution Function of Distribution value (DFDV). Therefore, for partition j one gets nj DFDV, j = 1,2,…,p. To include dependence of the p partitions, we construct a joint cummulative distribution as follows:
(4)
where
and
To model above joint cumulative distribution, we use copula model. A copula is a mathematical function that describes the dependence structure between random variables, enabling the modeling of their marginal distribution dependencies [37–41]. By utilizing the distribution functions which specify the distribution of distribution values at points
, and the joint distribution function HT, we can formulate a copula function for all
in the following manner:
(5)
where C denotes the copula function. In our classification problem, it is assume that the observation data are divided into k classes. Therefore, using the previous approaces, we obtain k copulas. The expression of the copula model for class k can be written in following form:
(6)
Let the random variables be defined as transformations
that following a uniform distribution on the interval [0,1]. The statistical relationship among these random variables can be expressed through a copula function in the following manner:
(7)
Furthermore, the relationship between the joint distribution function and the joint probability density function is expressed as follows:
(8)
With
denotes the joint probability density function of the random variables
. Subsequently,
represents the copula function that establishes the connection between the distribution of standardized random variables
and
. Additionally,
denotes the marginal distribution of each respective random variable.
This study employs the Gaussian copula model to characterize the relationships among random variables within images, specifically those associated with partitions and differentiating points. The Gaussian copula function utilized in this research is defined as follows:
(9)
with ΦΛ represents the multivariate standard normal distribution function with a mean vector of 0 and covariance matrix Λ. Additionally, Φ−1 denotes the inverse of the standard normal distribution.
The density function of the Gaussian copula can be derived from the Gaussian copula distribution function as follows:
(10)
With consists of standardized random variables, and |Λ| represents the determinant of the covariance matrix [41]. The joint density function of the Gaussian copula can be expressed as follows:
(11)
The joint density function (h) depicts the pattern of relationships among random variables associated with partitions and distinguishing points.
Parameter estimation
The Inference Function for Marginals (IFM) is frequently utilized in estimating copula model parameters. IFM aims to identify optimal values for the model parameters by maximizing the likelihood function, representing the probability of observing the data under the copula model [42]. This likelihood function is based on the joint probability density function of the data, influenced by the copula parameters. The formulation of the likelihood function is as follows:
(12)
with Λ represents the copula parameter, and Θ is the marginal parameter. In parameter estimation, the IFM often involves the use of the logarithm of the likelihood function [43]. The log-likelihood function for the copula is given as follows:
(13)
for the Gaussian copula, the likelihood function is given as follows:
(14)
with
.
The IFM method utilizes the marginal likelihood to estimate marginal parameters and the copula likelihood to estimate copula parameters. The approach involves a two-step process: first, estimating the marginal parameters and then using these estimates as inputs for the copula likelihood function to estimate copula parameters [40, 41]. The log-likelihood function of the Gaussian copula can be decomposed into two distinct components: the log-likelihood function component of the copula, given in Eq (15), and the log-likelihood function component of the marginal, given in Eq (16) as follows:
(15)
(16)
The optimization process is initiated by addressing the marginal components to derive the parameters of the marginal distribution [44–46]. The IFM algorithm is structured as follows:
Step 1: Determine the log-likelihood function for the Gaussian copula using Eq (14)
Step 2: Obtain the marginal parameter estimates by maximizing the log-likelihood function for the marginal, as given by Eq (16). For partitions j = 1,2,…,p, and points t = 1,2,…,nj:
(17)
Step 3: Substitute from step 2 to obtain the log-likelihood function for the Gaussian copula using Eq (15).
Step 4: Obtain the Gaussian copula parameter estimates by maximizing the log-likelihood function for Gaussian copula as follows:
(18)
Therefore, the parameter estimates using the IFM method are acquired by satisfying the following equations:
(19)
The procedure for estimating parameters in the marginal distribution is contingent upon the specific nature of the chosen distribution. Within the Gaussian copula model framework, the marginal distribution may be parametric or non-parametric. An exemplification of a parametric distribution is the beta distribution, defined by a domain function with values confined to the interval [0, 1]. The probability density function of the beta distribution for each j = 1,2,…,p, and t = 1,2,…,nj is expressed as follows:
(20)
Where and
, with
and
denoting the parameters of Beta distribution. These parameters are estimated through the Maximum Likelihood Estimation (MLE) method, utilizing the log-likelihood Beta function as follows:
(21)
Classification and evaluation
In this study, the employed classification method adopts a probabilistic framework [29], articulated as follows:
(22)
Here, represents the new input (testing), and Ck denotes the class of image k. Consequently,
signifies the conditional probability of class k given the input.
The term embodies the joint distribution modeled by the Gaussian copula. Additionally, P(Ck) denotes the proportion of training data for class k, computed as follows:
(24)
Where Nk training represents the number of training samples for class k, and N training signifies the total quantity of training data. Consequently, the conditional probability is expressed as:
(25)
After completing the classification process, the next step is model evaluation. Model evaluation ensures the model’s efficacy in predicting the image class. In this study, model evaluation is performed using the confusion matrix. Classification performance, in this context, is assessed using accuracy metrics. In addition, precision, recall, and F1 scores are used to comprehensively assess the model’s performance. Similar studies also use other metrics for classification evaluation [47]. These evaluation metrics measure image classification’s accuracy, completeness, and effectiveness.
The analysis steps of this study can be illustrated in a research flowchart that reflects the main steps. Initially, a pre-processing phase is carried out, including data cleaning, normalization, and splitting the data into training and testing sets. Next, the data preparation for training and testing is conducted, followed by the formation of ECDF and DFDV as random variables in the Gaussian copula model. Then, the formulation of the Gaussian copula model is executed to capture the dependencies among random variables, involving the estimation of model parameters. This estimation is obtained by applying the IFM method to achieve optimal estimates. The subsequent step involves the classification process using the developed model, followed by the evaluation and testing phase using specific evaluation metrics. It is important to note that the parameters used in the evaluation process of the testing data are the parameters obtained from the training model. The research flowchart is presented in Fig 1.
Illustrating stages and processes in image classification using Gaussian copula with symbolic data ECDF and DFDV.
Data characterization and pre-processing
This study introduces a Gaussian copula model for image classification. The data used in this study is MNIST handwritten numeric image data. The dataset consists of 60,000 grayscale images depicting numbers 0–9 for training and an additional 10,000 for testing. As part of the preprocessing step, the data undergoes normalization, ensuring each image is represented as a 20 × 20-pixel array to effectively capture the edges of handwritten numerals. Normalization ensures uniformity in the data, allowing the model to focus on the essential features of the numerals without being affected by variations in size or intensity. After normalization, the partitioning process was run, resulting in systematically organized partitions that included complete, two-line, three-line, and four-line data sets. This partitioning enhances the model’s ability to discern patterns and features within the handwritten numerical images visually depicted in Fig 2 as follows:
This study uses the full data and various partitions to investigate the effect of the number of partitions on the performance of the classification model. A larger number of partitions may lead to a scenario where there are too few pixels in each partition, thus prompting the restriction to four rows. Partitioning the data aims to analyze the impact of different levels of detail on the model’s accuracy. A key consideration involves striking a balance between achieving a more detailed image representation through partitioning and ensuring an adequate number of pixels for stable analysis. Too few pixels in a partition could lead to a loss of critical information, while too many partitions could introduce noise and complicate the classification task.
The process involves representing image features through a matrix of pixel intensities. To streamline data representation and analysis, the matrix of pixel values transforms into a vector, a procedure commonly known as vectorization. Adopting a row-by-row approach, each row of the pixel value matrix is systematically extracted and merged into a data vector, as outlined [48]. Consequently, each pixel value matrix is transformed into a corresponding data vector. This transformation facilitates the handling and processing of image data by converting it into a more manageable form for machine learning algorithms. When row-wise vectorization is applied to a matrix denoted as A20×20.
The elements of the data matrix A have components x, which is a representation of the pixel intensity values. We convert each x element of the data matrix, initially in the range [0,255], to the range [0,1] by dividing it by 255, expressed as . This transformation results in
values that are easy to handle in numerical calculations, especially in machine learning algorithms and statistical analyses, which are more stable and fast when performed in the interval [0,1]. When the vectorization process converts the data matrix A into a vector, the result is the vector Vec(A). Moreover, the transformation of matrix A yields the matrix
, which subsequently transforms into the vector
. Next, the ECDF is constructed by following the method specified in Eq (1). Fig 3 visually illustrates the construction of the ECDF. The ECDF serves as an image characteristic, capturing the pixel values that are less than or equal to x. These characteristics show variations in different image classes, underscoring the importance of identifying specific realization points of x that can effectively distinguish them. The ECDF [rovides a statistical summary of the image data, highlighting key features that contribute to the classification performance. The resulting vector
and ECDF
is visually illustrated in Fig 3:
Here, x.tilde represent , and F(x.tilde) represent
.
ECDF and DFDV results
Each image generates an ECDF with a total of N images, and a set of ECDF is produced for these images. At specific points, denoted as T1 and T2, the ECDF values from each function are gathered to construct a Distribution Function of Distribution Values (DFDV) as defined in Eq (2). This DFDV represents the distribution function of pixel ECDF values, illustrating the characteristics of pixel value utilization at points T1 and T2. These characteristics can vary across image classes, and selecting different points will result in different characteristics, thus referred to as differentiating points. In this case the specific distinguishing points selected are T1 = 0.1 and T2 = 0.7. The results of ECDF formation for each class in the full dataset are presented in Fig 4:
Here, x.tilde represent , and F(x.tilde) represent
.
The displayed image shows an ECDF plot of the value for digits 0 to 9 (k = 0 to k = 9) in the MNIST data set. Each subplot describes the cumulative distribution of normalized pixel values
to certain digits, from 0 to 9. Each subplot’s X axis displays a normalized pixel value, meaning it divides the original pixel value by 255 to fall within the range [0,1]. The Y axis indicates the cumulative value of the ECDF, which reflects the accumulative proportion of pixels with an intensity less than or equal to the value on the X axis. The red and blue vertical lines mark the specific distinguishing points selected are T1 = 0.1 and T2 = 0.7
The ECDF plot illustrates the distribution of normalized pixel intensity values for each digit in the MNIST data set. By looking at the variations and curve shapes in each subplot, we can obtain information about the distribution patterns of unique pixel strength values for every digit. This analysis is useful to understand how the pixel value distribution features of different digit images behave and can help in the development of accurate classification models.
The results of the ECDF analysis reveal distinctive patterns among classes, indicating variations in the characteristics of pixel intensity distributions across different classes. A notable disparity is observed in the ECDF for the digits 0 and 1. The distribution of the ECDF values at the differential points effectively represents the visible differences in distribution characteristics between these image classes. Significant differences in ECDF digits 0 and 1 indicate that the intensity of pixels and their distribution patterns are different, which can help distinguish between different digits. For example, a 0 digit may have a more uniform pixel distribution compared to a 1 digit, which tends to have greater intensity variations in some areas of the image. These discernible differences in distribution characteristics among image classes can be effectively represented by the distribution of ECDF values at differentiating points.
A DFDV is constructed at each differentiating point from the ECDF values set. This DFDV serves as the image characteristics for each partition and differentiating point. This DFDV effectively describes the probability distribution of image features with values less than or equal to a certain value. This DFDV is interconnected as it is derived from the same set of images. This interdependence is modeled in terms of Gaussian copulas, with the DFDV treated as random variables. The DFDV formation process and Gaussian copulas are visually depicted in Fig 5:
Formed DFDV Used as Random Variables in Gaussian Copula Model. Here, x.tilde represent , and F(x.tilde) represent
.
The interdependence of distribution values within a single image at differentiating points is modelled in this study using a copula, with and
serving as random variables. Various copula models, including Gaussian, Clayton, Frank, t-Student, and Gumbel, are available for consideration. The choice of the copula type is determined based on the scatter plot of the ECDF values at points T1 and T2 as random variables. The patterns observed in the scatter diagram are utilized to identify the most suitable copula model [48]. This approach employs graphical representations to discern the optimal copula model based on the observed relationships between random variables. The scatter plot results for
and
for classes k = 0,1,…,9 are presented in Fig 6:
Indicates Gaussian Copula Model with Linear and Elliptical Patterns. Here, x.tilde represent .
After selecting the Gaussian copula model, the next step is to specify the DFDV at points T1 and T2. These DFDV values capture the distribution characteristics at these points and are crucial for further analysis within the Gaussian copula framework. The DFDV results for classes k = 0,1,…,9 at T1 = 0.1 and T2 = 0.7 are shown in Fig 7.
Distinct distributions are evident among different classes.
The DFDV results for the image classes at the distinguishing points show different distributions. Visually, the separate shapes of these distributions indicate that the characteristics of the image classes are different. This highlights the ability of DFDV to capture and represent the unique features of each class, which is crucial for accurate classification and further analysis. By providing a clear differentiation between the classes, DFDV enhances the model’s ability to accurately identify and classify images based on their distinctive features.
Gaussian copula results
The Gaussian copula model applied to image classes k = 0,1,…,9, for the full dataset, specifically with j = 1 (only one partition) and differentiating points at t = 1,2, as described in Eq (8), is formulated as follows:
(26)
The term represents the random variable
, which follows a uniform distribution on the interval [0,1]. The notation Λ denotes the covariance matrix serving as parameter for Gaussian copula. The function ΦΛ represents the multivariate standard normal distribution, and Φ−1 denotes the inverse of the standard normal distribution. This model provides a mathematical representation that captures the dependency structure among various image classes, emphasizing the relationships between variables within the dataset.
The model presented in Eq (26) involves two random variables, and the density model of the Gaussian copula can be expressed as follows:
(27)
with
. This function characterizes the dependency structure between two specific random variables, influenced by the parameter matrix Λ. The Gaussian copula joint density function of Eq (27) can be given as follows:
(28)
The probability density function described here comprises two components. The initial component encapsulates the interdependence structure among the random variables within the copula function, while the subsequent component denotes the marginal distribution for each random variable. The copula function serves to characterize the collective behavior of the random variables, capturing their dependence structure. Parameter estimation for this model is undertaken through the Inference Functions for Margins (IFM) method. This approach utilizes the log-likelihood function, as delineated in Eq (13).
The parameter estimation process starts with the marginal component, which includes the distribution of each random variable. After obtaining the parameters for the marginal distribution, parameter estimation for the Gaussian copula component follows. In this study, the marginal distribution chosen is the beta distribution since the domain of the random variable is in the interval [0,1]. The results of the beta parameter estimation using Eq (16) are presented as Table 1 below:
Case k = 0,1,…,9, j = 1 and t = 1,2 for random variables and
.
Table 1 shows the beta parameter estimates for the Gaussian copula model for various image classes. Comparison between classes and observations at points T1 = 0.1 and T2 = 0.7 provides insight into the variability of the dependence and sensitivity of the model at those points. The beta parameter estimates provide information on the dependence structure and characteristics of the pixel intensity distribution.
Parameter estimation is conducted by maximizing the Gaussian copula log-likelihood function using the IFM method, as expressed in Eq (17). The estimation results of the covariance matrix parameters are obtained as Table 2 follows:
Case k = 0,1,…,9, j = 1 and t = 1,2 for random variables and
.
The covariance matrix values appear relatively high, indicating a strong relationship between the random variables and
. However, these values show minimal variation across classes, indicating that, with the full dataset, the characteristics that distinguish between classes cannot be adequately discerned. The introduction of image partitioning techniques becomes very important to improve the discrimination of object characteristics among image classes.
In the case of partitioning, the formation of ECDF and DFDV is performed for each partition. The random variables used in the Gaussian copula model depend on the number of partitions and discrimination points defined. For example, with two discrimination points, the complete data set model consists of two random variables. When the partitions are two, four random variables are involved, increasing to six for three partitions and eight for four partitions. Using the Gaussian copula model, as articulated in Eq (9) and the joint probability density model described in Eq (10), the Gaussian copula model is formulated for each scenario as Table 3 follows:
These Models are Applied to Image Class k = 0,1,…,9.
Classification and evaluation results
After obtaining the Gaussian copula model, characterized by the joint density function with marginal distribution parameters from Table 2 and Gaussian copula parameters from Table 3, the subsequent phase involves the classification process. The classification is executed through the probabilistic method outlined in Eqs (21–23). Before commencing the classification stage, the initial step involves preparing the training data, which includes pre-processing, image pixel extraction, data transformation, and ECDF formation. ECDF values are determined at the same distinguishing points utilized in the training process. The outcomes of ECDF testing at discrimination points T1 = 0.1 and T2 = 0.7 are presented as Fig 8 below:
Case j = 1 and T1 = 0.1 and T2 = 0.7.
Next is the classification process using the input ECDF values at points T1 = 0.1 and T2 = 0.7 for the training data. The classification results are given as Table 4 follows:
Case j = 1,2,3,4 and T1 = 0.1 and T2 = 0.7.
Table 4 presents the results of classification evaluation using the Gaussian copula model in various data partition scenarios. This evaluation includes several classification performance metrics, such as Accuracy, Precision, Recall, and F1-score. These results reflect the model’s performance when using the entire dataset without partitioning. The low accuracy indicates that the model needs help to classify diverse classes.
Dividing the data into two partitions significantly improves accuracy and other performance metrics. The model performs better in recognizing and separating features in both partitions. This improvement continues as the data is divided into three partitions, with accuracy and other performance metrics improving further. This suggests that using three partitions provides a better representation of the dataset’s variation.
With data divided into four partitions, there is a significant improvement in all performance metrics. The model achieves its highest performance with four partitions, indicating that more partitions provide a better representation and support more detailed feature separation. However, it should be noted that increasing the number of partitions beyond this might result in partitions without objects, causing instability in the formation of ECDF. Therefore, an increase in the number of partitions needs to be carefully managed to avoid instability in ECDF analysis.
The previous research utilizing the MNIST dataset for handwritten digits has implemented various classification techniques, including linear classification with error rates ranging from 7.6% to 12%, K-NN approach with error rates ranging from 1.5% to 5%, non-linear classification with an error rate of approximately 3.5%, and other methods such as SVM with error rates ranging from 1.6% to 4.7%. Additionally, CNN has been used with error rates between 0.7% and 1.7% [49]. The subsequent progress in research indicates a decrease in error rates. The advantage of the approach we applied is its simplicity and ease of interpretation. Therefore, the symbolic data approach using Gaussian copula can be considered a valid alternative for image classification.
Conclusion
Based on the analysis and experimental findings, it can be concluded that the Gaussian copula model, utilizing a cumulative distribution function space approach, stands as a viable and effective approach for conducting image classification on the MNIST dataset, demonstrating a high level of accuracy. The Gaussian copula proves its capability to model the interdependence of pixel usage across different image partitions. These partitions are strategically designed to encompass diverse object characteristics within the images. With an increase in the number of partitions, the model becomes adept at capturing finer distinctions in object characteristics, facilitating the differentiation between different numbers. The research successfully achieves its main objective of building a classification model rooted in the Gaussian copula, employing a cumulative distribution function space approach, thereby enhancing the clarity and efficiency of image classification.
Acknowledgments
We express our gratitude to everyone who contributed to this paper. Special thanks to the reviewers for their valuable input, and to our editorial team and peers for their support.
References
- 1.
Szeliski R. Computer Vision: Algorithms and Applications. Second. Switzerland: Springer; 2022. 1–937 p.
- 2. Druzhkov PN, Kustikova VD. A survey of deep learning methods and software tools for image classification and object detection. Pattern Recognit Image Anal. 2016;26(1):9–15. https://doi.org/10.1007/978-3-030-34372-9.
- 3. Yang MH, Kriegman DJ, Ahuja N. Detecting Faces In Image: A Survey. IEEE Trans Pattern Anal Mach Intell. 2002;24(1):1–25.
- 4. Gan J, Wang W, Lu K. A new perspective: Recognizing online handwritten Chinese characters via 1-dimensional CNN. Inf Sci (Ny). 2019;478:375–90. https://doi.org/10.1016/j.ins.2018.11.035.
- 5. Salma Shofia Rosyda, Tito Waluyo Purboyo. A Review of Various Handwriting Recognition Methods. Int J Appl Eng Res. 2018;13(2):1155–64.
- 6. Vashist PC, Pandey A, Tripathi A. A comparative study of handwriting recognition techniques. Proc Int Conf Comput Autom Knowl Manag ICCAKM 2020. 2020;456–61.
- 7. Rajalakshmi M, Saranya P, Shanmugavadivu P. Pattern Recognition-Recognition of Handwritten Document Using Convolutional Neural Networks. IEEE Int Conf Intell Tech Control Optim Signal Process INCOS 2019. 2019;1–7.
- 8. Assegie TA, Nair PS. Handwritten digits recognition with decision tree classification: A machine learning approach. Int J Electr Comput Eng. 2019;9(5):4446–51. https://doi.org/10.11591/ijece.v9i5.pp4446-4451.
- 9. Yaswanth M, Raj NI, Nath MK. Hopfield Neural Network for Classification of Digits. 9th IEEE Uttar Pradesh Sect Int Conf Electr Electron Comput Eng UPCON 2022. 2022;9986401. https://doi.org/10.1109/UPCON56432.2022.9986401.
- 10. Yamashita R, Nishio M, Do RKG TK. Convolutional neural networks: an overview and application in radiology. Insights Imaging. 2018;9:611–29. pmid:29934920
- 11. Ahmed Medjahed S. A Comparative Study of Feature Extraction Methods in Images Classification. Int J Image, Graph Signal Process. 2015;7(3):16–23. https://doi.org/10.5815/ijigsp.2015.03.03.
- 12. Zhang H, Berg AC, Maire M, Malik J. SVM-KNN: Discriminative nearest neighbor classification for visual category recognition. Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2006;2:2129–36.
- 13. Kumar A, Sminchisescu C. Support Kernel machines for object recognition. Proc IEEE Int Conf Comput Vis. 2007;
- 14. Zhang Y, Wu L, Neggaz N, Wang S, Wei G. Remote-sensing image classification based on an improved probabilistic neural network. Sensors. 2009;9(9):7516–39. pmid:22400006
- 15. Thai LH, Hai TS, Thuy NT. Image Classification using Support Vector Machine and Artificial Neural Network. Int J Inf Technol Comput Sci. 2012;4(5):32–8. https://doi.org/ 10.5815/ijitcs.2012.05.05.
- 16. Mohapatra RK, Majhi B, Jena SK. Classification performance analysis of MNIST Dataset utilizing a Multi-resolution Technique. 2015 Int Conf Comput Commun Secur ICCCS 2015. 2016;1–5.
- 17. Li Q, Cai W, Wang X, Zhou Y, Feng DD, Chen M. Medical image classification with convolutional neural network. 2014 13th Int Conf Control Autom Robot Vision, ICARCV 2014. 2014;2014(December):844–8.
- 18. Kuo CCJ. Understanding convolutional neural networks with a mathematical model. J Vis Commun Image Represent. 2016;41:406–13.
- 19. Shijie J, Ping W, Peiyi J, Siping H. Research on data augmentation for image classification based on convolution neural networks. Proc—2017 Chinese Autom Congr CAC 2017. 2017;2017-Janua(201602118):4165–70.
- 20. Paoletti ME, Haut JM, Plaza J, Plaza A. A new deep convolutional neural network for fast hyperspectral image classification. ISPRS J Photogramm Remote Sens [Internet]. 2018;145:120–47. Available from: https://doi.org/10.1016/j.isprsjprs.2017.11.021
- 21. Sultana F, Sufian A, Dutta P. Advancements in image classification using convolutional neural network. Proc—2018 4th IEEE Int Conf Res Comput Intell Commun Networks, ICRCICN 2018. 2018;122–9.
- 22. Wang W, Yang Y. Development of convolutional neural network and its application in image classification: a survey. Opt Eng. 2019;58(04):1–20.
- 23. Ashraf R, Habib MA, Akram M, Latif MA, Malik MSA, Awais M, et al. Deep Convolution Neural Network for Big Data Medical Image Classification. IEEE Access. 2020;8:105659–70.
- 24. Gong R, Tang Z. Further investigation of convolutional neural networks applied in computational electromagnetism under physics-informed consideration. IET Electr Power Appl. 2022;16(6):653–74.
- 25. Xin M, Wang Y. Research on image classification model based on deep convolution neural network. Eurasia J Image Video Process. 2019;2019(1).
- 26. Stitou Y, eddine Lasmar N, Berthoumieu , Stitou Y, eddine Lasmar N, Berthoumieu Y Copulas based multivariate Gamma modeling for texture classification To cite this version: HAL Id: hal-00727108 COPULAS BASED MULTIVARIATE GAMMA MODELING FOR TEXTURE. 2012;
- 27. Salinas-Gutiérrez R, Hernández-Aguirre A, Rivera-Meraz MJJ, Villa-Diharce ER. Using Gaussian Copulas in supervised probabilistic classification. Stud Comput Intell. 2010;318:355–72.
- 28. Bansal R, Hao X, Liu J, Peterson BS. Using Copula distributions to support more accurate imaging-based diagnostic classifiers for neuropsychiatric disorders. Magn Reson Imaging. 2014;32(9):1102–13. pmid:25093634
- 29. Sen S, Diawara N, Iftekharuddin KM. Statistical Pattern Recognition Using Gaussian Copula. J Stat Theory Pract. 2015;9(4):768–77.
- 30. Tamborrino C, Mazzia F. Classification of hyperspectral images with copulas. J Comput Math Data Sci [Internet]. 2023;6(March 2022):100070. Available from: https://doi.org/10.1016/j.jcmds.2022.100070
- 31. Lasmar NE, Berthoumieu Y. Gaussian copula multivariate modeling for texture image retrieval using wavelet transforms. IEEE Trans Image Process. 2014;23(5):2246–61. pmid:24686281
- 32. Li C, Huang Y, Xue Y. Dependence structure of Gabor wavelets based on copula for face recognition. Expert Syst Appl [Internet]. 2019;137:453–70. Available from: https://doi.org/10.1016/j.eswa.2019.05.034
- 33. Kuiry S, Das N, Das A, Nasipuri M. EDC3: Ensemble of Deep-Classifiers using Class-specific Copula functions to Improve Semantic Image Segmentation. 2020;1–12. Available from: http://arxiv.org/abs/2003.05710
- 34. Winarni S, Indratno SW, Sari KN. Character of images development on gaussian copula model using distribution of cumulative distribution function. Commun Math Biol Neurosci. 2021;2021:1–13.
- 35. Winarni S, Indratno SW, Sari KN. Pemodelan Gambar Menggunakan Copula Gaussian Dengan Metode Partisi. Stat J Theor Stat Its Appl. 2021;21(1):37–43.
- 36. Winarni S, Indratno SW, Arisanti R, Pontoh RS. Image Feature Extraction Using Symbolic Data of Cumulative Distribution Functions. 2024;
- 37. Vrac M, Chédin A, Diday E. Clustering a global field of atmospheric profiles by mixture decomposition of copulas. J Atmos Ocean Technol. 2005;22(10):1445–59.
- 38. Schweizer B. Introduction to Copulas. J Hydrol Eng. 2007;12(4):346–346.
- 39. Nelsen RB. Properties and applications of copulas: A brief survey. First Brazilian Conf Stat Model Insur Financ [Internet]. 2003;3:1–18. Available from: http://matematica.campusnet.unito.it/didattica/att/d11d.7507.file.pdf
- 40. Durante F, Puccetti G, Scherer M, Vanduffel S. My introduction to copulas. Vol. 5, Dependence Modeling. 2017. 88–98 p.
- 41. Joe H. Dependence modeling with copulas. Dependence Modeling with Copulas. 2014. 1–457 p.
- 42. Trivedi PK, Zimmer DM. Copula modeling: An introduction for practitioners. Vol. 1, Foundations and Trends in Econometrics. 2005. 1–111 p.
- 43. Kim G, Silvapulle MJ, Silvapulle P. Comparison of semiparametric and parametric methods for estimating copulas. Comput Stat Data Anal. 2007;51(6):2836–50.
- 44. Choroś B, Ibragimov R, Permiakova E. Copula Estimation. 2010;77–91.
- 45. Ko V, Hjort NL. Model robust inference with two-stage maximum likelihood estimation for copulas. J Multivar Anal [Internet]. 2019;171:362–81. Available from: https://doi.org/10.1016/j.jmva.2019.01.004
- 46. Bickel EP, Diggle P, Fienberg S, Gather U. Copula Theory and Its Applications. 2009.
- 47. Anbalagan T, Nath MK, Vijayalakshmi D, Anbalagan A. Analysis of various techniques for ECG signal in healthcare, past, present, and future. Biomed Eng Adv [Internet]. 2023;6(January):100089. Available from: https://doi.org/10.1016/j.bea.2023.100089
- 48. Nixon MS, Aguado AS. Feature extraction and image processing for computer vision. Academic Press Elsevier; 2019. 1–626 p.
- 49. Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–324.