Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Leaf identification using radial basis function neural networks and SSA based support vector machine

  • Ali Ahmed,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Visualization, Writing – original draft

    Affiliation IT Department, Faculty of Computers and Information, Menoufia University, Menoufia, Egypt

  • Sherif E. Hussein

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – review & editing

    sherif_hussein@mans.edu.eg

    Affiliation Computer Engineering and Systems Department, Faculty of Engineering, Mansoura University, Mansoura, Egypt

Abstract

In this research, an efficient scheme to identify leaf types is proposed. In that scheme, the leaf boundary points are fitted in a continuous contour using Radial Basis Function Neural Networks (RBFNN) to calculate the centroid of the leaf shape. Afterwards, the distances between predetermined points and the centroid were computed and normalized. In addition, the time complexity of the features’ extraction algorithm was calculated. The merit of this scheme is objects’ independence to translation, rotation and scaling. Moreover, different classification techniques were evaluated against the leaf shape features. Those techniques included two of the most commonly used classification methods; RBFNN and SVM that were evaluated and compared with other researches that used complex features extraction algorithms with much higher dimensionality. Furthermore, a third classification method with an optimization technique for the SVM using Salp Swarm Algorithm (SSA) was utilized showing a significant improvement over RBFNN and SVM.

I. Introduction

Nowadays, plants play an essential role in our life. They are beneficial for human food, medicine and industry. The identification of each species is a difficult task because of the existence of numerous plant species and the similarity between them. Automation and identification of plant species are extremely important for scientists interested in many fields including agriculture and botany.

The identification and recognition of plant species have been carried out by researchers for many years utilizing different plant features such as leaf shape, texture, and color. Different plant species have different features that provide beneficial information for researchers. One category of these features, commonly used in literature, is called histograms of oriented gradients (HOGs) that are applied in many visual object recognition applications [1]. While other research directions used parse representation of leaf tooth features for automatic pieces identification that include the number, rate, sharpness, and obliqueness concatenated to form a feature vector [2].

In another investigation, several geometric features (12 morphological features) were defined in [3] and divided into two different categories, namely general and domain-related visual features. A research by Yanikoglu et al. has utilized a general category of features that consists of color, texture, and shape features [4]. In addition, Jyotismita et al, characterized and recognized plant’s leaves by extracting both texture-based and shape-based features [5]. They applied Gabor filter and Gray-Level Co-occurrence Matrix to model leaf’s texture and calculate a set of Curvelet transform coefficients and invariant moments to represent the leaf’s shape. Another research direction used contour-based features such as centroid contour distance (CCD) and angle code (AC) that are considered the most popular features for many researchers [6]. The work presented in [7] extracted fifteen features from plant’s leaf identification. Among these features, eight geometric metrics such as rectangularity, circularity, eccentricity features together with moment invariant features. Moreover, an image-based description and measurement of the leaf’s teeth are used in [8].

In literature, various approaches for plant species identification and recognition are proposed by researchers. A deep learning scheme for the quantification and discrimination of plant leaf was presented in [9]. In their work, Convolutional Neural Networks (CNN) were used for learning useful leaf features that can be extracted directly from a dataset. Other researchers used a neuro-fuzzy controller (NFC) and a feed-forward back-propagation Multi-Layered Perceptron (MLP) to differentiate between thirty-one leaves’ classes [5]. Another type of classifiers, the Support Vector Machine (SVM) was trained with a portion of the training data using a subset of features which included color, shape, and texture features [4]. In a research presented in [10], a classification using k-Nearest Neighbor classifier was implemented and tested on 640 leaves belonging to 32 different species of plants based on a feature extraction stage. To distinguish leaf margins that cannot be easily differentiated by most of the commonly used geometric features, linear discriminant classifier (LDC) was utilized for leaves classification [11]. While another classification method called Moving Center Hypersphere (MCH) classifier was efficiently used to address high-dimensional leaf’s features [12]. In addition, the study presented in [13] used SVM for segmenting and classifying legume leaves based on the analysis of their veins. Tan et al. used a classifier, namely CNN for plant identification from leaf vein features [14]. They used the CNN classifier to discriminate between three types of legume species named white, red, and soybean. Moreover, a simple nearest neighbor classifier was used to classify plant leaves using texture features extracted by a modified local binary patterns (MLBP) [15]. Also, a simple linear discriminant (LDA) analysis and SVM were used in [16] to classify plant leaves. Table 1 summaries the related studies which are beneficial to compare the different identification techniques’ performance and their corresponding features.

In this research, two of the most successfully used classification techniques were presented and evaluated against the leaf shape features, namely RBFNN and SVM. Then, an optimization technique for the SVM using a Salp Swarm Algorithm (SSA) was proposed and compared with RBFNN and SVM. This research is organized as follows: Section two discusses in detail the different classification techniques used in this research; Section three presents the proposed leaf features extraction, Section four shows the results and discussion of the proposed work, and finally, Section five gives the conclusions for the proposed study and the suggestion for future research.

II. Methods

The introduced feature extraction algorithm in this research uses the leaf shape to identify the geometrical features. Furthermore, the proposed approach will be assessed using three classification methods and compared with the performance of other techniques found in recent literature that identify the different plant species. The utilized classification techniques are discussed in the following subsections:

A) RBFNN classification

Artificial neural network (ANN) can be incorporated in plenty of applications, including curve fitting, regression, and classification. Curve fitting is used to get the values for a set of parameters through which a given set of explanatory variables can depict another variable such as equation estimation using empirical methods. Regression, on the other hand, is unambiguously defined from the statistics perspective. In a sense, regression is not a mathematical concept; it's a statistical concept, while curve fitting is a mathematical concept. Regression is curve fitting under context: and the goal is to understand how one variable depends on another variable.

Radial basis function neural network (RBFNN) is the most commonly used approach for many research fields such as object classification, linear regression, curve fitting, and discrete-based data clustering [17]. It utilizes radial basis function as its activation function as in [18, 19]. RBFNN framework consists of three layers; input layer, hidden layer, and output layer. The first layer of RBFNN is the input vector to each unit in the next hidden layer. RBF activates each unit in the hidden layer. Then, a linear combination of the activations using each unit of the hidden layer is used to create a classified output. The output layer mainly depends on the activation function used and their weights combined with the links between both the hidden and the output layers. RBFNNs were used for both the feature extraction stage as well as the classification of the different leaf's types.

For the RBFNN as a classifier the input vector of the RBFNN architecture is an n-dimensional vector that represented the extracted features. Each neuron in the hidden layer stores one of the training set vectors and computes the similarity between the input features vector and the corresponding prototype vector and according to this comparison, the output layers give a value between 0 and 1. If the output unit of a certain RBF neuron is close to 1, then the input is equal to the prototype. If the difference between the input unit and the prototype increases, the output unit value falls towards zero exponentially. The shape of the RBF response of neurons takes a bell-shaped curve and the value of neuron's response value is the activation of the neuron. RBF output layer consists of different nodes; each node represents a class and computes a value for its corresponding class. The classification decision is done by assigning the input vector to the class that achieves the highest score as shown in Fig 1.

There are different similarity functions; Gaussian function is considered the most popular one. The following equation is a one-dimensional Gaussian function. (1) where x in Eq (1) is the input vector, μ is the computed mean value, and σ represent the standard-deviation. The activation function of RBF neuron is different from Gaussian function and is written as: (2) this function is the prototype vector that represent the center of the bell-shaped curve.

B) Support vector machine classification

SVM method is considered one of the commonly used machine learning methods for classification and regression problems and was first introduced by Drucker [20]. The SVM classifier is used to classify the unknown sample by building the classification model from the training data. The classifier has two parameters called penalty and kernel parameters. These parameters have major effect on SVM performance [21]. The first one plays an important role as a tradeoff between training data error minimization and the maximization of the classification margin. The other is used to determine the nonlinear mapping between both the input and the high-dimensional feature space. Consequently, the way to choose the values of the above-mention parameters affects the performance of SVM [22, 23].

The mis-classification might result in because of non-separable data. So, the linear SVM constraints should be slowed down and that can be done by adding a slack variable ξi as shown in the following equation (Eq (3)). The slack variable ξi represents the distance between each training sample and the corresponding margin which should be minimized.

(3)

These equations can be unified to one equation as follows: (4) where the slack variable ξi≥0

If 0 ≤ 0 ξi ≤ 1, then the sample is correctly classified as it lies between the classification margin and the correct side of the hyper plane. On the other hand, if ξi > 1, then, Yi(wTXi+B)≥1−ξi; therefore, the function (wTXi+B) and the label (Yi) have different sign and accordingly, the sample data Xi is mis-classified.

After adding the variable ξi and the penalty parameter C to the objective function of SVM, we obtain the following equation: (5)

By formalizing Eq (5) using Lagrange formula, it will appear as follows: (6) where αi in the equation is greater than zero.

After differentiating Lagrange formula LP with respect to the three variables w, B and ξi we obtain the following equation after setting the result to zero: (7)

It is clear from the above equation that αi determines the upper bound of the variable C. In addition, the support vectors (SVs) with the variable αi = C can lie inside or outside the margin boundary.

In case the sample is non-linearly separable data, SVM maps it to a linearly separable, as a higher-dimensional data, using the kernel function Q. According to Eq (8), this kernel is represented as a dot product of the nonlinear functions. The objective function of SVM is defined by Eq (9).

(8)(9)

In literature, there are several kernel functions used for mapping the data. Some of these are the linear kernel function defined as K(XI,Xj) = 〈XI,Xj〉 and RBF is defined as .

C) Salp swarm algorithm

Salp Swarm Algorithm (SSA) is a random population-based algorithm introduced by [24] and recently used by [25, 26]. The idea behind SSA is to simulate the swarming behavior of salps when foraging in oceans. They usually shape a swarm as a salp chain when foraging. In the proposed SSA, the leader salp is the front of the chain and the followers are the remaining of salps chain. According to the implementation of other swarm-based techniques, the salps locations in the chain are represented by a two-dimensional matrix. In addition, the target of swarm is assumed to be the food source called F. The following is the mathematical model of Salp swarm algorithm:

In SSA implementation, the following proposed equation is used to update the leader’s position as follows: (10) where is defined as the first leader’s position in jth dimension search space and Fj is the food source position in jth dimension, uppbjis the upper-bound and lowbj is the lower-bound of jth dimension respectively, the two parameters s2, and s3 are random numbers. As seen from Eq (10), the leader’s position of salp chain is updated according to the food source position. The first parameter s1 has its importance in SSA because of its role of balancing between exploration and exploitation as follows: (11) where the parameters Lc and LM in Eq (11) are defined as the current and the maximum numbers of iterations, respectively. The parameters s2 and s3 are uniformly generated from the interval [0, 1].

By increasing iteration count, s1 parameter decreases. That, in turn, can put more emphasize on the diversification on initial stages and put more emphasize on the intensification tendency at the last steps of optimization. The location of followers is adjusted by the following equation [24]: (12)

In this equation, i≥2 and the position represent the ith follower’s position in jth dimension space. The following algorithm is the pseudo code of SSA.

Algorithm 1. Pseudo code for the SSA.

Initialize pi (i = 1,2,…,n) the position of salp′s population and consider uppb and lowb

While (the termination criteria are not met)

                 Compute all search agent’s fitness (salps)

                 F is the food source (the best search agent)

                 Update s1 according to Eq ()

                 For each salp’s position (pi)

                                  If (i equal to 1)

                                  Update the leading salp’s location using Eq ()

                                  else

                                  Update the follower salp’s location using Eq ()

                                  end

                 end

                 modify the salps based on uppb and lowb

end

return F

D) The SSA-SVM classifier

In this sub-section, the proposed SSA-SVM model in the framework is described to compute the optimal SVM parameters as follows:

  • Initialization: the following parameters of SSA were initialized (salps number, maximum iterations, salp’s position and velocity).
  • Training: in the SSA-SVM model, the SSA inputs the SVM classifier by the two values, penalty parameter C and RBF kernel parameter, to train SVM. Therefore, each salp position is represented a combination of two parameters (C and kernel parameters) in two-dimensional search space.
  • Testing: the testing samples is used to compute the mis-classification (error rate) of each position dividing the number of mis-classified samples (Nerr) by the total number of testing samples (Ntotal) according to Eq (13). When the values of the two above-mentioned parameters at certain position pi∈R2 reached the minimum error rate, the optimal solution is located at pi∈R2,
(13)
  • Termination: When the end condition is reached, the process terminated; if not, the next generation operation proceeds. In our model, the SSA is ended when the operation reached the max iterations, or the obtained best solution is not changed within a certain number of iterations. In this case, the salp positions are then updated according to (Eqs 10 and 12).

III. Leaf features extraction

The proposed feature extraction algorithm uses the plant leaf shape, as illustrated in Fig 2, and is explained in detail as follows.

A) Image shape signatures

Shape signature is one of the most commonly used approaches to represent contour descriptors. It represents the shape using a one-dimensional function calculated from the contour points of the leaf's shape. There are plenty of shape signatures used in the literature. One of the most frequently used shape signatures is the centroid contour distance (CCD) according to [27, 28, 29, 30]. The studies in [31, 32, 33], calculated centroid contour distance plot and angle code histogram (ACH) for leaf representation. The authors in [34] represented the leaf shape using centroid contour distance curve and proposed an ontology-based integrated approach for classification. The extracted features points from the boundary are then stored in an ordered data structure such as linked list. Invariant moments are used to represent leaf’s shape.

The centroid of a discrete-based shape is computed using the following equations: (14) (15)

Where N is referring to the number of image’s boundary points. If we define the coordinate set of contour points of the shape as:

By considering the centroid of the leaf shape as a central reference point, the distance Dc between the centroid and each feature point can be calculated using the following equation.

(16)

To make the computed distances scaling-invariant, they were divided by the maximum distance and stored in a circular list.

RBFNNs are used in this stage for regression. They are applied as a curve fitting technique on the dataset to get the leaf contour. In which RBFNN parameters were adjusted to find the continuous relation between angle and contour distance using GA optimization. Those parameters include the total number of units in the hidden layer, RBF centers and standard deviations to minimize the RMSE below 1% for the testing data. The maximum number of generations for GA was set to 1000.

The dataset for each number of features were randomly arranged. The features representing the dataset are divided into training and testing datasets. One way to evaluate the classifier model is to split the training dataset into training and validation datasets, then train the models against the smaller training dataset and evaluate them against the validation dataset using k-fold cross-validation. We randomly split the training dataset into 10 distinct subsets called folds, followed by training and evaluating the classifier model 10 times, using a different fold for evaluation every time and training on the other 9 folds. The result is an array that contains the 10 evaluation scores. Moreover, cross-validation allows to get not only an estimate of the performance of the model, but also a measure of how precise this estimate is. As the dataset was medium-sized, 80% of the dataset features were used for training and validation, while 20% of the dataset were used for testing.

B) CCD sequence from shape boundaries

The information is captured from the shape’ boundary using the description of shape sequences. The appropriate formulation of centroid contour distances using RBFNN curve fitting as well as its properties according to scale, translation and rotation invariance is discussed. The centroid contour distance (CCD) sequence represents the distance between each contour boundary point and the contour centroid. One of the characteristics of CCD curve sequence is its translation invariance as well as scale invariance under certain normalization.

To prove the scale invariance of the proposed scheme, a binary leaf image was reduced to different sizes. Then, the corresponding curve fitting is created using the features of these images as shown in Fig 3. The features used are the normalized distances as measured relative to the maximum distance between the centroid and the farthest point on the contour. Therefore, the scale effect will vanish by the normalization process. Therefore, the normalized features are scale invariant. Also, to prove the rotation invariance, a binary leaf image was rotated anti-clockwise to different degrees and then their features were extracted from the rotated images to compute their corresponding curve fitting as shown in Fig 4. Because the features depend on the maximum distance between the centroid and the farthest point on the contour as the reference axis, that means features are rotation invariant as the reference axis will be the same in any rotated leaf. An example of declaring that different plant leaves be members of the same species may contain great differences in their shape contour as shown in Fig 5. This figure shows that two plant leaves belong to the same species have different feature curves.

thumbnail
Fig 3.

A binary leaf image reduced to different sizes and their corresponding curve-fitted lines to verify scale invariance. The horizontal axis of curve-fitted lines represents the angle’ values and the vertical axis of the curve-fitted lines, represent their corresponding normalized CCD values.

https://doi.org/10.1371/journal.pone.0237645.g003

thumbnail
Fig 4. Binary leaf image rotated anti-clock wise from the origin degree to 40o, 80o, 120o, and 160o from left to right respectively to verify rotation invariance with their corresponding curve-fitted data.

The horizontal axis of the curve-fitted lines represents the angle values, and the vertical axis of the curve-fitted lines represents the corresponding normalized CCD values.

https://doi.org/10.1371/journal.pone.0237645.g004

thumbnail
Fig 5. Two different leaf images belonging to the same plant species and their corresponding contour feature curves.

https://doi.org/10.1371/journal.pone.0237645.g005

IV. Results and discussion

This section evaluates the performance of recognizing the various plant species. The utilized classification techniques, namely RBFNN, SVM and an optimized version of the SVM using a metaheuristic optimization technique known as Salp Swam Algorithm will be applied and compared using the proposed leaf features extraction.

A) Dataset

In this research, we used plant leaves dataset that publicly available from UCI [35, 36]. The dataset contains leaves images for about 100 plant species. Each one includes sixteen data shapes. Each data shape is characterized by 64 features vectors obtained from the normalized contour signatures. These signatures were calculated from the differentiation of shape image’s foreground and background pixel.

B) RBFNN based contour fitting

The centroid for the contour points for each leaf was calculated, and the contour points’ distances to that centroid were normalized relative to the maximum distance. Those normalized distances represented the dependent variable while the angles of those lines were calculated relative to the maximum contour distance to the centroid as the independent variable. RBFNN was then used to find the best curve that minimizes the RMSE for each leaf in the dataset. The performance of that curve fitting is shown in Fig 6 in which several leaf examples were illustrated. For the RBFNN, the optimized curves along with other curves with different spread values, were shown to demonstrate the effect of the parameter on the fitting accuracy. Fig 7 illustrates the different classes plotted according to different pairs of features. The graphs in two dimensions show the complexity of separating the leaf’s classes and the necessity for the classifiers to perform well in the complex and high dimensional problems.

thumbnail
Fig 6. Different leaf species with the corresponding Mean Square Error (MSE) and RBFNN curve fitting.

RBFNN curve fitting.

https://doi.org/10.1371/journal.pone.0237645.g006

thumbnail
Fig 7. Pairs of features from the 36 features defining the leaf and their corresponding leaf classes.

https://doi.org/10.1371/journal.pone.0237645.g007

C) RBFNN classifier

The number of features points for each leaf were selected to be of different sizes. Those features are uniformly distributed starting from angle 0o relative to the maximum distance between the contour and the centroid and is incremented by a constant angle to achieve the required number of features. Those same features will be used by the other two classifiers to ultimately choose the number of features that corresponds to the best performance. The dataset features were arranged randomly while 80% of the dataset were used for testing and validation using 10-fold cross-validation while 20% were used for testing. The choice of the ratio of 80:20 for training and testing was done because of the medium-sized dataset.

The Genetic algorithm was used here to tune the RBFNN parameters (weights, μ, and β) to achieve minimum RMSE with a population size of 20 solutions and 1000 number of generations.

D) SVM classifier

SVM classifier was applied on the dataset using the Gaussian kernel. However, the minor increase of the value of σ, will increase the fitting for the training data on the expense of the generalization error and can cause over-fitting problem. While the regularization parameter C compromises between smooth decision boundary and correctly classifying the training points. Therefore, to achieve better classification accuracy, these parameters were optimized by minimizing the mis-classification rate on the datasets and accordingly ensure high classification performance for the leaf dataset using Bayesian optimization approach. The training curve shown in Fig 8 illustrates the steep descent of the objective function value with the number of iterations (function evaluation). While Fig 9 shows the kernel parameters optimization using Bayesian optimization technique. It is worth mentioning here that SVM is very similar to RBFNN and can achieve comparable results. However, the excessive number of parameters used by RBFNN can be reflected in the algorithm complexity, and that can, in turn, lead sometimes to a lower performance if the parameter selection is not optimized.

thumbnail
Fig 8. The SVM objective function minimization against iterations.

https://doi.org/10.1371/journal.pone.0237645.g008

thumbnail
Fig 9. The SVM kernel parameters optimization using Bayesian optimization.

https://doi.org/10.1371/journal.pone.0237645.g009

E) SSA based SVM classifier

The salps’ positions are first initialized using random numbers. The searching limit of the parameter C of the support vector machine (SVM) was limited by Cmin = 0.01 and Cmax = 1000, and the searching limit of σ was determined by σmin = 0.01 σmax = 100. If these limits increase, it will enlarge the search space; and accordingly, more salps will need to be utilized to search for the optimal solution and that will lead to more computation and consequently a slow convergence rate.

In this research, the number of search agents is 20, while the algorithm is iterated until the objective function value of 0.025 is met. Fig 10 shows the performance curve (minimum objective function achieved) for the algorithm plotted against the iterations during the training stage. While Fig 11 shows the kernel parameter optimization using the SSA optimization technique. The SSA based SVM gives a slightly faster curve to reach the objective function than in the conventional SVM as shown in Fig 8.

thumbnail
Fig 10. The SSA based SVM objective function minimization against iterations.

https://doi.org/10.1371/journal.pone.0237645.g010

thumbnail
Fig 11. The SVM kernel parameters optimization using SSA optimization.

https://doi.org/10.1371/journal.pone.0237645.g011

In order to find the best features, different angles were selected relative to the leaves reference axes. The experiments started with an angle of 30o which corresponds to 12 points on the contour representing the leaf’s features. The three proposed classifiers have been applied using the dataset of all the leaves’ features and the classification accuracies were calculated. The experiments were then repeated for angles 20o, 10o, and 8o which correspond to numbers of features 18, 36, and 45 respectively. The results achieved by the three classifiers are summarized in Table 2.

thumbnail
Table 2. The proposed system accuracies with respect to different selected angles (number of features).

https://doi.org/10.1371/journal.pone.0237645.t002

As inferred from Table 2, the best performance of the proposed classifiers has resulted from using a 36 features dataset. Therefore, the three corresponding classifiers were further evaluated. Performance metrics comparison is summarized in Table 3, in which the processing time for each technique was evaluated and broke down into the different task times consumed during the training and validation stage. In all the classification techniques evaluated in this research, k-fold cross-validation was used in which data were partitioned to nearly k subsets with the same approximate sizes. The platform used in this research to develop the three classification techniques is implemented using a PC with Intel(R) core (TM) i7-7500 CPU, 2.7 GHz, 16 GB RAM, Windows 10 operating system, and MATLAB (R2019b).

thumbnail
Table 3. Classification parameters of the proposed three classifiers.

https://doi.org/10.1371/journal.pone.0237645.t003

The relative performance of consumed time and classification accuracies showed that RBFNN consumes more time to reach the objective function value (0.025) and at the same time with a classification accuracy of 91.45%. While SVM classifier managed to achieve the shortest consumed time with a relatively high classification accuracy of 93.33%, which can compare well with the other findings in the literature. On the other hand, SSA based SVM managed to achieve a high classification accuracy of 96.67% on the expenses of slightly longer time consumption. Therefore, the SSA based SVM combined with the RBFNN feature extractor based on the shape of the leaf could offer a promising paradigm in similar applications.

V. Conclusions

RBFNN is widely used in complex applications with a great accuracy in both regression and classification. It has been used here to find the best fitting curve for each leaf contour data points to allow for any selection scenario of features. That can be important in similar problems that rely on the object shape and is able to avoid the effect of rotation, scaling, and translation. The RBFNN was used again in the classification stage with considerable classification accuracy compared to the other methods found in the literature. On the other hand, when SVM was used with the selected features, an accuracy of 93% was reached, which outperformed the RBFNN classifier. The parameters of the RBF kernel used in the SVM classifier were the key for further SVM accuracy improvement. Therefore, SSA, a recent optimization technique that belongs to the meta-heuristic techniques that proved to be superior in avoiding local optima due to their stochastic nature, was considered. It was modified in a way to be slower when it is too close to the optimal point. That was done due to the fact it is a quasi-optimal technique that never guarantees the solution reached to be the optimal one. When the SSA was used with SVM, it reached 96.67, which represent a significant performance increase compared to the RBFNN classifier and the SVM classifier. That shows the potential in many similar fields that might benefit greatly from a features’ extractor using RBFNN and a classifier using the SSA based SVM. Moreover, a future research direction is to investigate other applications that incorporate geometric features in image processing, especially those with irregular shapes, which can elaborate on the proposed feature extraction approach.

References

  1. 1. Olsen A., Han S., Calvert B., Ridd P., & Kenny O. In Situ Leaf Classification Using Histograms of Oriented Gradients. International Conference on Digital Image Computing. Techniques and Applications (DICTA), 2015.
  2. 2. Jin T., Hou X., Li P., & Zhou F. A Novel Method of Automatic Plant Species Identification Using Sparse Representation of Leaf Tooth Features. PLOS ONE, 2015, Volume 10, No. 10, e0139482. pmid:26440281
  3. 3. Zulkifli Z., Saad P., & Mohtar I. A. Plant leaf identification using moment invariants & General Regression Neural Network. 11th International Conference on Hybrid Intelligent Systems (HIS), 2011, pp. 430–435.
  4. 4. Yanikoglu B., Aptoula E., & Tirkaz C. Automatic plant identification from photographs. Machine Vision and Applications, 2014, Volume 225, No. 6, pp. 1369–1383.
  5. 5. Chaki Jyotismita, Parekh Ranjan, Bhattacharya Samar. Plant Leaf Recognition using Texture and Shape features with Neural Classifiers. Pattern Recognition Letters, 2015, Volume 58, pp. 61–68.
  6. 6. Fotopoulou F., Laskaris N., Economou G., & Fotopoulos S. Advanced leaf image retrieval via Multidimensional Embedding Sequence Similarity (MESS) method. Pattern Analysis and Applications, 2011, Volume 16, No. 3, pp. 381–392.
  7. 7. Wang X., Du J., & Zhang G. Recognition of Leaf Images Based on Shape Features Using a Hypersphere Classifier. Lecture Notes in Computer Science, 2005, pp. 87–96.
  8. 8. Kamilaris A. and Prenafeta-Boldú F., 2018. Deep learning in agriculture: A survey. Computers and Electronics in Agriculture, 147, pp.70–90.
  9. 9. Lee S. H., Chan C. S., Mayo S. J., & Remagnino P. How deep learning extracts and learns leaf features for plant classification. Pattern Recognition, 2017, Volume 71, pp. 1–13.
  10. 10. Munisami T., Ramsurn M., Kishnah S., & Pudaruth S. Plant Leaf Recognition Using Shape Features and Colour Histogram with K-nearest Neighbour Classifiers. Procedia Computer Science, 2015, Volume 58, pp. 740–747.
  11. 11. Kalyoncu C., & Toygar Ö. Geometric leaf classification. Computer Vision and Image Understanding, 2015, Volume 133, pp. 102–109.
  12. 12. Wang X., Huang D., Du J., Xu H., & Heutte L. Classification of plant leaf images with complicated background. Applied Mathematics and Computation, 2008, Volume 205, No. 2, pp. 916–926.
  13. 13. Wang H, Tian D, Li C, Tian Y, Zhou HPlant leaf tooth feature extraction. PLoS ONE, 2019, Volume 14, No. 2. e0204714. pmid:30759085
  14. 14. Tan J. W., Chang S., Binti Abdul Kareem S., Yap H. J., & Yong K. Deep Learning for Plant Species Classification using Leaf Vein Morphometric. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 2018, 1–1.
  15. 15. Grinblat G. L., Uzal L. C., Larese M. G., & Granitto P. M. Deep learning for plant identification using vein morphological patterns. Computers and Electronics in Agriculture, 2016, Volume 127, pp. 418–424.
  16. 16. Larese M. G., Namías R., Craviotto R. M., Arango M. R., Gallo C., & Granitto P. M. Automatic classification of legumes using leaf vein image features. Pattern Recognition, 2014, Volume 47, No. 1, pp. 158–168.
  17. 17. Moawad N., Elawady W. and Sarhan A., 2019. Development of an adaptive radial basis function neural network estimator-based continuous sliding mode control for uncertain nonlinear systems. ISA Transactions, 87, pp.200–216.
  18. 18. Abbassi R., Abbassi A., Heidari A., & Mirjalili S. An efficient salp swarm-inspired algorithm for parameters identification of photovoltaic cell models. Energy Conversion and Management, 2019, Volume 179, pp. 362–372.
  19. 19. Mirjalili S. (2018). Evolutionary Radial Basis Function Networks. Studies In Computational Intelligence, 105–139. https://doi.org/10.1007/978-3-319-93025-1_8
  20. 20. Miller S. and Startz R., 2019. Feasible generalized least squares using support vector regression. Economics Letters, 175, pp.28–31.
  21. 21. Al-Zoubi A., Faris H., Alqatawna J., & Hassonah M. (2018). Evolving Support Vector Machines using Whale Optimization Algorithm for spam profiles detection on online social networks in different lingual contexts. Knowledge-Based Systems, 153, 91–104.
  22. 22. Deng W., Yao R., Zhao H., Yang X., & Li G. (2017). A novel intelligent diagnosis method using optimal LS-SVM with improved PSO algorithm. Soft Computing, 23(7), 2445–2462.
  23. 23. Gao L., Ye M., Lu X., & Huang D. (2017). Hybrid Method Based on Information Gain and Support Vector Machine for Gene Selection in Cancer Classification. Genomics, Proteomics & Bioinformatics, 15(6), 389–395. pmid:29246519
  24. 24. Mirjalili S., Gandomi A., Mirjalili S., Saremi S., Faris H., & Mirjalili S. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Advances in Engineering Software, 2017, Volume 114, pp.163–191.
  25. 25. Ibrahim A., Ahmed A., Hussein S., & Hassanien A.: Fish Image Segmentation Using Salp Swarm Algorithm. The International Conference On Advanced Machine Learning Technologies And Applications (AMLTA2018), 2018, pp. 42–51.
  26. 26. Zareapoor M., Shamsolmoali P., Kumar Jain D., Wang H., & Yang J. (2018). Kernelized support vector machine with deep learning: An efficient approach for extreme multiclass dataset. Pattern Recognition Letters, 115, 4–13.
  27. 27. Fachri Yanuar Rudi F., & Yuniarno E. M. (2019). Contour to centroid distance graph as feature in hand gesture recognition. IOP Conference Series: Materials Science and Engineering, 536, 012150.
  28. 28. Wäldchen J., & Mäder P. (2017). Plant Species Identification Using Computer Vision Techniques: A Systematic Literature Review. Archives Of Computational Methods In Engineering, 25(2), 507–543. pmid:29962832
  29. 29. Wang S., Kim H., & Chen H. Contour Tracking Using Centroid Distance Signature and Dynamic Programming Method. 2009 International Joint Conference on Computational Sciences and Optimization, 2009, pp. 274–278.
  30. 30. Yanikoglu B., Aptoula E., & Tirkaz C. (2014). Automatic plant identification from photographs. Machine Vision And Applications, 25(6), 1369–1383.
  31. 31. Sokic E., & Konjicija S. (2016). Phase preserving Fourier descriptor for shape-based image retrieval. Signal Processing: Image Communication, 40, 82–96.
  32. 32. Wu M., Xiao W., & Hong Z. (2018). Similar image retrieval in large-scale trademark databases based on regional and boundary fusion feature. PLOS ONE, 13(11), e0205002. pmid:30439950
  33. 33. Chang C., Hwang S., & Buehrer D. A shape recognition scheme based on relative distances of feature points from the centroid. Pattern Recognition, 1991, Volume 24, No. 11, pp. 1053–1063.
  34. 34. Khmag A., Al-Haddad S. A., & Kamarudin N. (2017). Recognition system for leaf images based on its leaf contour and centroid. 2017 IEEE 15th Student Conference on Research and Development (SCOReD).
  35. 35. Dua D. and Graff C. (2019). UCI Machine Learning Repository [http://archive.ics.uci.edu/ml]. Irvine, CA: University of California, School of Information and Computer Science.)
  36. 36. Frank A., Asuncion A. UCI machine learning repository, Online 2010. https://archive.ics.uci.edu/ml/index.php