Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A new Multi Sine-Cosine algorithm for unconstrained optimization problems

  • Muhammad Zubair Rehman,

    Roles Conceptualization, Methodology, Writing – original draft

    Affiliation Soft Computing & Data Mining Centre (SMC), Faculty of Computer Science & Information Technology (FSKTM), Universiti Tun Hussein Onn Malaysia, Parit Raja, Malaysia

  • Abdullah Khan ,

    Roles Formal analysis, Funding acquisition, Writing – original draft, Writing – review & editing

    abdullah_khan@aup.edu.pk

    Affiliation Institute of Computer Sciences and IT (ICS/IT), The University of Agriculture, Peshawar, Pakistan

  • Rozaida Ghazali,

    Roles Data curation, Funding acquisition, Project administration

    Affiliation Soft Computing & Data Mining Centre (SMC), Faculty of Computer Science & Information Technology (FSKTM), Universiti Tun Hussein Onn Malaysia, Parit Raja, Malaysia

  • Muhammad Aamir,

    Roles Data curation, Investigation, Methodology

    Affiliation Soft Computing & Data Mining Centre (SMC), Faculty of Computer Science & Information Technology (FSKTM), Universiti Tun Hussein Onn Malaysia, Parit Raja, Malaysia

  • Nazri Mohd Nawi

    Roles Conceptualization, Project administration, Resources, Software, Supervision, Validation, Writing – original draft

    Affiliation Soft Computing & Data Mining Centre (SMC), Faculty of Computer Science & Information Technology (FSKTM), Universiti Tun Hussein Onn Malaysia, Parit Raja, Malaysia

Abstract

The Sine-Cosine algorithm (SCA) is a population-based metaheuristic algorithm utilizing sine and cosine functions to perform search. To enable the search process, SCA incorporates several search parameters. But sometimes, these parameters make the search in SCA vulnerable to local minima/maxima. To overcome this problem, a new Multi Sine-Cosine algorithm (MSCA) is proposed in this paper. MSCA utilizes multiple swarm clusters to diversify & intensify the search in-order to avoid the local minima/maxima problem. Secondly, during update MSCA also checks for better search clusters that offer convergence to global minima effectively. To assess its performance, we tested the MSCA on unimodal, multimodal and composite benchmark functions taken from the literature. Experimental results reveal that the MSCA is statistically superior with regards to convergence as compared to recent state-of-the-art metaheuristic algorithms, including the original SCA.

Introduction

In modern times, optimization has become pertinent to the development of reliable and robust solutions in the field of science and engineering. Optimization involves certain searching mechanisms that can find the best solutions against an objective function [1]. Mostly, optimization search techniques are categorized into deterministic and stochastic search algorithms. The deterministic algorithms mostly use gradient descent trajectories and extremas to find the most feasible solution. Meanwhile, stochastic search uses multiple points in search space and finds multiple optimal solutions in the search space with more efficacy than the deterministic algorithms [2]. Metaheuristic search mostly employs the stochastic search mechanisms that result in the provision of efficient solutions to non-linear problems [3]. A metaheuristic improvement is grounded in the belief that a stochastic estimate of an optimum will be better than a deterministic solution [2]. Due to their prowess at finding optimal solutions with less computing power, several metaheuristics inspired by nature and physical phenomena have been developed. Some classical examples are differential evolution (DE) [4], genetic algorithm (GA) [5], particle swarm optimization (PSO) [6], ant colony optimization (ACO) [7], cuckoo search (CS) [8], wolf search (WS) [9], artificial bee colony (ABC) [10], bat algorithm (BA) [11], harmony search (HS) [12], and simulated annealing (SA) [13] etc. More recently, a new set of nature-inspired metaheuristics or simply swarm intelligent metaheuristics are developed to solve large-scale optimization problems. These algorithms are grey wolf optimizer (GWO) [14], crow search algorithm (CSA) [15], African buffalo optimization (ABO) [16, 17], whale optimization (WOA) [18], and Sooty-Tern (STOA) [19] etc. The goal of all metaheuristic algorithms is to keep a fine balance between exploration & exploitation of the search space [20]. Although a significant number of algorithms are proposed in this field successfully, but no free lunch (NFL) theorem still welcomes more newbies by suggesting that all algorithms perform optimally in the right environment [3, 21]. Satisfied with NFL theory, Mirjalili et al. proposed a Sine-Cosine algorithm (SCA) in 2015 that uses Sine and Cosine functions for improved metaheuristic search [22]. Despite showing its prowess at solving several optimization problems, it still has problems like low learning curves. Also, its magnitude changes progressively during each iteration that ensues in Sine-Cosine’s early commitment to exploitation which can be counter-productive. Also, in case of many local optimum, Sine-Cosine can converge to a sub-optimal solution [3]. To address the issues faced by the Sine-Cosine algorithm (SCA), this paper proposes an improved Multi Sine Cosine Algorithm (MSCA) that will avoid local optima convergence and improves exploitation. This characteristic makes MSCA suitable for solving optimization problems with multiple local minimums because it maintains a balance between exploration and exploitation with clustered population of solutions. To assess its performance, we adopt the proposed MSCA for the optimization benchmark functions suite in [22]. MSCA’s performance was evaluated against eight popular metaheuristic algorithms, including artificial bee colony (ABC) [23], butterfly optimization algorithm (BOA) [24], crow search algorithm (CSA) [15], differential evolution (DE) [4], grasshopper optimization algorithm (GOA) [25], harmony search (HS) [26], salp swarm optimization (SSA) [27], and the SCA. Experimental results reveal that the MSCA exhibits competitive performance as compared to the SCA and other eight meta-heuristic algorithms. Our contributions are summarized as follows:

  • A new Multi sine-cosine algorithm (MSCA) that permits the selection of local and global search operations. In the first stage, MSCA offers clustered population to diversify & intensify the search in order to avoid the local minima. Secondly, during the update, MSCA also checks for better search clusters that offer convergence to global minima effectively. Then, these clusters are merged to form a single cluster X that offers a better solution during the search.
  • MSCA is tested on 19 complex functions used by Mirjalili in [22]. The optimization problems in these experiments include unimodal, multimodal and composite functions of both low/high dimensions.
  • MSCA performed efficiently on hard optimization problems when compared with the state-of-the-art metaheuristic algorithms. Moreover, the efficiency of multiple population division concept of MSCA inspires us to investigate its effect on other metaheuristic algorithms in future studies.

The structure of the paper is organized as follows: Section 2 shed some light on the literature review in which it shed some light on the previous improvements on SCA algorithm; then the original SCA algorithm is discussed in the Sections 3. Section 4 describes the proposed MSCA algorithm. Result and discussions are presented in Section 5. Finally, the paper is concluded in the Section 6.

Literature review

Owing to the theory of “No free lunch theorem” [28] that makes it impossible for a single algorithm to be a remedy for all optimization problems, Sine-Cosine algorithm (SCA) was proposed in 2016 [22]. SCA works by utilizing the sine and cosine functions to generate a set of interdependent candidate solutions. The new position of the solutions is highly dependent on the previous candidate solution. Since its inception, SCA has been widely utilized on single objective benchmark functions testing [22] and multi-objective functions [29]. Besides benchmark functions, SCA has been applied for solving unit commitment problem in energy production [30] and in feedforward neural networks to predict the liver enzymes of carp fish with high accuracy [31]. Hafez et al. [32] applied SCA for binary feature selection and minimization and to enhance the classification performance. Not satisfied with the convergence in SCA, Elaziz et al. stated that it gets stuck in local minima because the operators used for exploration do not work well [21]. He successfully augmented the performance of SCA with an opposition-based learning platform to generate better solutions. Meshkat et al. came up with an idea of a new weighted update position mechanism (WUPM) instead of the original update method of search agents in SCA. In this method, each search agent was assigned a weight based on its fitness and the position of each agent is updated based on the previous weighted position of the search agent [33]. In 2017, SCA was applied for clustering binarized images of handwritten Arabic text with less noise [34] and for optimization of space shuttle trajectory [35]. A new trend of hybridization was seen at the end of 2017 with the introduction of different high-level heuristics (HLL) and low-level heuristics (LLH) algorithms in SCA. The examples are SCA with Differential Evolution (DE) for structural damage assessment of a truss [36], SCA with Grey Wolf Optimizer (GWO) [37] & SCA with Crow Search for optimization functions [38], Adaptive SCA with Particle swarm optimization (ASCA-PSO) for pairwise local sequence alignment [39], SCA with whale optimization for parameter optimization in a milling process [40], hybrid self-adaptive sine cosine algorithm with opposition based learning [41], and improved sine–cosine algorithm based on orthogonal parallel information for global optimization [42] etc. To find a fine balance between exploration and exploitation, levy flight [43] and chaotic maps [44] were integrated with SCA. Zamli et al. [3] used reinforcement Q-learning table to maintain the states of Sine-Cosine in 2018. They used a reward and punishment mechanism to switch between levy flight and the crossover operator to enhance the solution’s diversity. End of 2018 saw a considerable rise of SCA getting integrated with machine learning techniques such as; parameter enhancement of support vector machines [45], a binary variant of SCA [46], context-based image segmentation [47], breast cancer classification [48], secure data placement in the Internet of Things [49], image thresholding [50], load frequency control of autonomous power system using adaptive fuzzy based PID controller optimized with improved sine cosine algorithm [51] etc.

Later years brought several noteworthy contributions towards the existing SCA algorithm to improve its convergence properties. Most of the proposed improvements were on the modification of exploration and exploitation strategies in SCA. Guo et al., adopted optimal neighborhood and quadratic interpolation strategy to overcome the problem of population update that is guided by the global optimal state in the SCA algorithm. The proposed QISCA used a stochastic optimal neighbor for neighborhood updates, and a quadratic interpolation curve for individual updates. Also, the population’s exploration was enhanced with quasi-opposition learning strategies thus improving the convergence speed and accuracy [52].

In order to improve the exploitation ability of the SCA, a symmetric SCA with adaptive probability selection (SSCA-APS) was introduced. The proposed SSCA searched normally in the early stages using the default parameters. In the later stages, it dynamically adjusted step-sizes of the search with adaptive probability selection (i.e., to integrate original and symmetric sine-cosine operators). Gaussian distribution was used to avoid the local minima by mutating the global optimal individuals in the current generation. A new individual in population is achieved through quasi-interpolation of two randomly selected individuals with a global optimal individual. SSCA-APS was considered a better improvement when tested on benchmark test functions against other SCA variants [53]. Same year, a multi-strategy enhanced SCA algorithm was proposed by Chen et al. to overcome the problem of local optima in SCA for large dimensional problems. The proposed SCA variant employed five strategies (i.e., Cauchy mutation operator, chaotic local search mechanism, opposition-based learning strategy and two operators based on differential evolution) to converge to global optima while maintaining a fine balance between exploration and exploitation. The performance of the proposed SCA variant was verified against other variants on several CEC2014 benchmark and real functions. The simulation results showed that the proposed SCA is better in terms of quality of solutions and convergence speed [54].

Optimal design of off-grid and on-grid hybrid energy management and supply is quite challenging when the energy production is mostly relying on changing climatic conditions. Renewable energy generated from wind turbines are one of the effective carbon emission control strategy of this age. But wind energy’s availability is stochastic in nature. To overcome this problem, Guesmi et al. proposed the integration of chaotic Sine-Cosine algorithm to improve the exploration and exploitation problem of SCA. The improved CSCA algorithm was able to minimize the economic emission dispatch efficiently when applied on the 69-bus ten-unit and 40-unit test systems [55]. Similarly, in another study an improved Sine-cosine with inertial weight algorithm (ISCA) is proposed to design optimal energy management systems for hybrid photovoltaic/wind/fuel cell (PV/WT/FC) system. The target was to minimize the cost of hybrid system life span (CHSLS) for a remote in Iran. Simulation results showed that the ISCA finds easily the optimal combination as PV/WT/FC system with minimal CHSLS than PSO and SCA algorithms [56].

Hydropower systems are another source of clean energy and optimizing operations in hydropower reservoirs can not only help in increased power utilization but also maintain optimal water distribution in arid agricultural zones. Feng et al., tried to answer the reservoir problem with adaptive SCA algorithm. ASCA algorithm used elite mutation strategy to overcome population diversity problem in SCA and the simplex dynamic search strategy to improve solution’s quality in real-time hydropower operations in China [57]. In a similar study by Feng et al., three stage optimization strategy is employed to improve SCA. Quasi-opposition learning strategy is used to find a fine balance between exploration and exploitation. The adaptive mutation is employed in the same manner as in ASCA algorithm. And finally, a random weighting agent generated by multiple leader solutions is integrated into the agent’s evolution equation to improve the overall convergence rate of SCA. The proposed SCA algorithm was tested on several benchmark composite functions. Also, this method showed its mettle in terms of quality of solution and convergence rate when tested on long-term reservoir operations in China [58].

Modern power transmission networks are becoming quite complex due to the integration of several distributed generators (DG). Directional overcurrent relays (DOCR) protect such networks in a highly constrained environment. Sarwagya et al., employed SCA to solve the optimal coordination problems of DOCR on faults generated by 3-bus, 8-bus, 15-bus and 30-bus test systems. SCA was found to effectively reduce the coordination interval time between primary and backup relays [59]. In another similar study, Raut and Mishra proposed multi-objective sine-cosine algorithm for optimal DG allocation. The objectives were to optimally allocate DG in radial distribution systems while minimizing total active power loss, annual energy loss cost, pollutant gas emissions, and maximizing voltage stability index. Their proposed approach was found to be effective; when, it was tested on slightly larger 33-bus and 69-bus distribution systems under four practical load conditions against strength Pareto evolutionary algorithm 2, non-dominated sorting genetic algorithm II, and multi-objective particle swarm optimization [60]. In another study on DG, SCA was successfully used to reduce the effect of global warming and environmental pollution due to the fossil fuel thermal energy generation by integrating wind energy in hydrothermal scheduling (HTS) [61].

With ever growing size of datasets, feature selection offers an efficient way to reduce a dataset’s dimensions and extracting useful information. For this purpose, a multi-objective sine cosine algorithm (SCA) for feature selection (MOSCA_FS) is proposed for hyperspectral imagery. MOSCA_FS is modeled to minimize the redundancy and maximize the relevance of the selected features. MOSCA_FS is found to an effective framework when tested on several benchmark hyperspectral image datasets [62]. To enhance the response of accuracy and response time in machine learning, a new variant called Improved Followers of Salp swarm Algorithm using Sine Cosine algorithm and Disrupt Operator (ISSAFD) is proposed. The ISSAFD works by updating the position of the followers in SSA using sine-cosine algorithm which helps in avoiding local optima altogether [63]. In another high-level hybridization, ABC was effectively integrated with SCA to optimize the threshold values during image segmentation for reduced search region [64]. In the late 2020, SCA was also successfully utilized for reducing features in text categorization in bag of words model [65], and for optimizing local optima problem in Volleyball premier league (VPL) algorithm with high level hybridization [66].

Earlier in 2021, Lawal et al. proposed sine cosine algorithm with artificial neural network (SCA-ANN) models for predicting blast-initiated ground vibrations in five granite mines. The goal was to maximize the safety of human lives and properties near the mining zones by minimizing the impact of vibrations emanating from the blasting of rocks. The proposed SCA-ANN was tested against the Gene expression programming (GEP), adaptive neuro-fuzzy inference system (ANFIS) for predicting the peak particle velocity (PPV). SCA-ANN model trained on 100 datasets and was found to be better than other models with an accuracy of 99 percent with almost zero error [67]. Seeing the intrinsic noise and outlier problems occurring during data clustering in the real world due to initial centroid selection, Kuo et al., proposed a robust SCA-FPCOM algorithm. The proposed SCA-FPCOM originated from the combination of probability c-means, fuzzy c-ordered means, and the sine-Cosine algorithm to improve clustering. The SCA-FPCOM was tested on several datasets and was found to be better than conventional approaches in-terms of adjusted rand index and the Silhouette coefficient [68].

Solar power on-grid generation is becoming a norm these days to avoid generating large electricity bills by selling excess electricity to the grid. Therefore, it is necessary to forecast solar power generation accurately in different weathers. Dash et al., proposed a hybrid forecasting approach consisting of empirical wavelet transform (EWT), Robust minimum variance Random Vector Functional Link Network (RRVFLN) optimized with Sine-Cosine algorithm. The proposed EWT-RRVFLN-SCA was found to be better than the original RRVFLN methods when tested on the historical solar power data [69]. In the mid of 2021, Hussain et al. proposed a hybrid Sine-Cosine Harris Hawk optimization (SCHHO) for minimum feature selection and maximum model generation. Their proposed SCHHO algorithm integrated since-cosine with Harris Hawk optimization to eliminate ineffective exploration in HHO. The SCHHO algorithm was able to adjust exploitation in HHO with dynamic adjustment of candidate solutions thus avoiding stagnancy in local solutions. The SCHHO was tested against state-of-the-art hybrid algorithms on sixteen datasets with high-dimensions exceeding 15000 attributes for numerical optimization. The proposed SCHHO was able to increase convergence speed and reduced features to 87 percent and achieved an accuracy of up to 92 percent [70]. Timeline of some of the most significant contributions to SCA are given in the Fig 1.

thumbnail
Fig 1. Some of the major contributions on SCA since its inception.

https://doi.org/10.1371/journal.pone.0255269.g001

Sine-Cosine algorithm

The Sine-Cosine algorithm (SCA) is a metaheuristic search algorithm proposed by Mirjalili et al. [22]. SCA works by searching solutions in the search based on the sine or cosine function given in Eqs (1) or (2) respectively: (1) (2)

In general, Both Eqs (1) and (2) are combined into one function as given in the Eq (3): (3)

Where Pi is the destination solution, Xi is the current solution, || indicates the absolute value. r1, r2, r3 and r4 are the random variables. The parameter r1 is a random variable responsible for determining the area of the next solution, this area may be either outside space between Xi and Pi or inside them. Mirjalili et al. [22] update the parameter r1 using the following Equation to balance exploration and exploitation. The effect of r1 can be seen in the Fig 2.

(4)

Where, a is a constant, T is the maximum number of iterations and t is the current iteration.

thumbnail
Fig 2. The decreasing pattern of Sine-Cosine indicating a switch from exploration to exploitation in the search space [71].

https://doi.org/10.1371/journal.pone.0255269.g002

The r2 is a random variable which used to find the direction of the movement of the next solution (i.e., if it is towards or outwards Pi). Also, the r3 is a random variable which gives random weights for Pi to stochastically emphasize (r3>1) or de-emphasize (r3<1) the effect of destination in defining the distance. The r4 is used to switch between the sine and cosine functions as in Eq (3). The steps of the Sine-Cosine algorithm are given in the Algorithm 1 (Fig 3).

Improved Multi Sine-Cosine Algorithm (MSCA)

The proposed MSCA algorithm enhances the SCA algorithm in two stages; the first stage offers a clustered population to diversify & intensify the search to avoid the local minima. Secondly, during the update, MSCA also checks for better clusters that offer convergence to global minima effectively. The proposed MSCA algorithm starts by generating random clusters of search agent’s i.e., X1, X2, X3,…,Xn of equal sized population N, in which each cluster Xi = [xi1, xi2, xi3,…,xin] represents a complete solution to the specified problem. Then, these clusters are merged together to form a single group X that offers a better solution. The steps of the algorithm are as follows;

  1. Initialize the random population clusters, X1, X2, X3,…,Xn.
  2. The value of r1 in Eq (3) plays a pivot role in MSCA to decide whether to switch from explore or exploit. It gradually decreases from 2 to 0.
  3. If, the value of r1>1, then, MSCA tries to offer better diversity in the new single solution. It merges all the clusters using maximum Euclidean distance (maxED) in Eq (5);
    (5)
  4. If, the value of r1<1, then, MSCA tries to offer better intensity in the newly merged single solution by using minimum Euclidean distance (minED) in Eq (6);
    (6)
  5. Evaluate each of the search agent clusters by the Objective function (f(X))
  6. Update the Best solution out of all clusters Obtained so far (P = X)
  7. Update r1, r2, r3, and r4
  8. Update the position of search agents using Eqs (5 and 6)
  9. Continue evaluating against the Objective function until all the conditions are met or the iterations are finished

The proposed Multi Sine-Cosine algorithm is demonstrated in Fig 4.

Results and discussions

In this section, MSCA is tested on benchmark functions and its performance is compared with eight algorithms. The detailed implementation, parameter settings, and results are discussed in the sub-sections.

Implementation of MSCA and the benchmark algorithms

The proposed MSCA algorithm’s performance is evaluated on the benchmark functions commonly used by Mirjalili et al. [71]. The system used for simulations was an Intel Core i5 processor with 8GB of RAM. The proposed MSCA was implemented on MATLAB R2018b with Windows 10 and compared with the latest and most popular algorithms given in Table 1. Three types of benchmark functions were used to evaluate the efficiency of the proposed MSCA algorithm, i.e., unimodal, multimodal, and composite. The detailed descriptions of the function and their properties are given in the sub-sections.

thumbnail
Table 1. Initialization parameter values for the comparison algorithms.

https://doi.org/10.1371/journal.pone.0255269.t001

Parameter settings

During all the experiments, default parameters were used for all the parallel algorithms. Meanwhile, the proposed MSCA algorithm used a population size of 30 that was equally divided among all six clusters. For a fair comparison, the maximum iterations for all the algorithms were set to 1000 with 30 trials on each function. The parameter setting of all the algorithms used in this paper is given in Table 1.

Statistical analysis

For statistical analysis, Standard deviation (SD) is used to find any variations in the average trial values and Mean formula is used to calculate the average of all trials. The equations of SD and Mean are expressed as [2]; (7) (8)

Where n is the total number of inputs, N is the total number of values or elements, xi is the number of input patterns, and x′ is the mean of xi.

Sometimes, it is not fair to say that the proposed algorithm is better because it performed statistically better in terms of mean and SD over 30 independent trials as SD and mean does not compare each run. Therefore, to decide on the significance of each result, the Wilcoxon rank-sum is used to find whether the proposed algorithm is statistically better in performance than the rest of the algorithms [72]. The null hypothesis H0 shows that there is no significant difference as far as the sample size is concerned for MSCA and each comparison algorithm. Alternative hypothesis, H1 means that MSCA’s sample size is less than that of each comparison algorithm [13].

Description of the benchmark functions

The proposed MSCA algorithm is evaluated based on the 19 mathematical functions used by Mirjalili et al. [71]. For testing the algorithms, the benchmark functions are divided into three complex types; i.e., Unimodal, multimodal, and composite functions etc. The two-dimensional (2-D) view of the functions is illustrated in the Fig 5.

thumbnail
Fig 5. Two-dimensional (2-D) view of the benchmark functions.

https://doi.org/10.1371/journal.pone.0255269.g005

The detailed descriptions of the function and their properties are given in Tables 2 and 3. Dimension, Range, and fmin columns in the Tables 2 and 3 denote the dimensions of the function, boundary of the search space, and the cost of the function.

thumbnail
Table 2. Mathematical formulae of the benchmark functions.

https://doi.org/10.1371/journal.pone.0255269.t002

Parametric test analysis

Depending on the complexity of the benchmark functions, they are divided into three types; i.e., Unimodal (F01-F06), multimodal (F07-F13), and composite (F14-F19). Unimodal functions have no local optima, a single global optimum and used to evaluate the exploitation capability of the algorithm. Multimodal functions have multiple local optima, a single global optimum and used to measure the exploration capability of the algorithm. Meanwhile, the composite functions are used to find the fine balance between exploration and exploitation capability of an algorithm. The statistical mean, SD and CPU time are given in the Tables 46 respectively.

thumbnail
Table 4. Average mean values by algorithms on benchmark functions.

https://doi.org/10.1371/journal.pone.0255269.t004

thumbnail
Table 5. Standard deviation (SD) of algorithms on benchmark functions.

https://doi.org/10.1371/journal.pone.0255269.t005

thumbnail
Table 6. Elapsed time of algorithms on benchmark functions.

https://doi.org/10.1371/journal.pone.0255269.t006

From Table 5, we can witness that the MSCA gained the best mean values on F1-F3, F7-F11, F13-F15, and F17-F19. Meanwhile, the second-best mean values were observed for ABC on F01- F04, F11, F13, F14, and F16-F18. Third best mean values were recorded on SCA for F01-F04, F07-F11, and BOA for F03-F04, F06, F08, F10, F15, F18. Best values were shown by BOA for F04, HS for F05, and DE for F06. Overall, it can be seen that MSCA has high searching precision on unimodal and multimodal problems. Despite the complexity they offer during search by composite functions, MSCA performed better. The CSA was found to be the most unstable algorithm in the above batch and was not able to perform on multimodal or composite functions. The mean convergence performance of all the algorithms can be seen in the Figs 6 and 7.

thumbnail
Fig 6. Convergence performance of MSCA algorithm on unimodal and multimodal functions.

https://doi.org/10.1371/journal.pone.0255269.g006

thumbnail
Fig 7. Convergence performance of MSCA algorithm on multimodal and composite functions.

https://doi.org/10.1371/journal.pone.0255269.g007

For standard deviation (SD), MSCA is able to perform with best SD values on F01-F03, F07, F09-F10, F13-F14, and F16-F18. Similarly, BOA showed best results on F04-F05, DE on F06, F11-F12, F14-F16, and HS on F08, F14, F16. Although, most of the algorithms showed stable SD on majority of the functions but ABC was the most stable among them. Again, worst deviation in results were shown by CSA. All SD values can be seen in the Table 6.

In recent age, CPU time is becoming negligible because of the ever-increasing speed of CPU’s but still it was recorded for all algorithm and results are illustrated in the Table 6. Here, BOA took the best average time on most functions i.e., F01-F03, and F09. CSA took the crown of best average time on most of the functions i.e., F04-F07, and F10-F11, followed by GOA which only performed on F13-F19, SCA on F12, and the proposed MSCA on F08.

Non-parametric test analysis

In this paper, the Wilcoxon rank-sum test is used to determine the significance of the results obtained by MSCA with five other algorithms. Table 7 shows the p-values obtained by MSCA and SCA. The superior significance of the proposed MSCA can be seen in Table 7, where it outperforms SCA, SSA, HS, GOA, and DE algorithms.

Conclusions

A new Multi Sine Cosine algorithm (MSCA) is introduced in this paper. The original Sine Cosine algorithm had a problem of early commitment to exploitation that leads it towards sub-optimal solutions. The proposed MSCA adopts the method of population clusters to diversify & intensify the search in order to avoid the local minima. Secondly, during the update, MSCA also checks for better search clusters that offer convergence to global minima effectively. The proposed MSCA avoids the premature convergence to local optima and avoids variance during convergence. MSCA’s performance is evaluated against eight popular metaheuristic algorithms, including ABC, BOA, CSA, DE, GOA, HS, SSA, and the SCA on 19 complex benchmark functions. Experimental results reveal that the MSCA exhibits competitive performance as compared to the SCA and other eight meta-heuristic algorithms. MSCA performed efficiently on hard optimization problems when compared with the state-of-the-art metaheuristic algorithms. Moreover, the efficiency of multiple population division concept of MSCA inspires us to investigate its effect on other metaheuristic algorithms in the future studies.

Supporting information

S1 Dataset. Benchmark functions/ datasets used in this paper.

https://doi.org/10.1371/journal.pone.0255269.s001

(DOCX)

References

  1. 1. Alamri HS, Zamli KZ. PMT: Opposition-Based Learning Technique for Enhancing Meta-Heuristic Performance. IEEE Access. 2019;7: 97653–97672.
  2. 2. Nawi NM, Rehman MZ, Khan A, Chiroma H, Herawan T. A modified bat algorithm based on Gaussian distribution for solving optimization problem. J Comput Theor Nanosci. 2016;13.
  3. 3. Zamli KZ, Din F, Ahmed BS, Bures M. A hybrid Q-learning sine-cosine-based strategy for addressing the combinatorial test suite minimization problem. PLoS One. 2018;13: 1–29. pmid:29771918
  4. 4. Price K V. Differential Evolution. Intell Syst Ref Libr. 2013.
  5. 5. Mirjalili S. Genetic algorithm. Studies in Computational Intelligence. 2019.
  6. 6. Kennedy J, Eberhart R. Particle swarm optimization. Neural Networks, 1995 Proceedings, IEEE International Conference on. 1995. pp. 1942–1948 vol.4. https://doi.org/10.1109/ICNN.1995.488968
  7. 7. Chniter M, Abid A, Kallel I. Towards a bio-inspired ACO approach for building collaborative learning teams. 2018 17th Int Conf Inf Technol Based High Educ Training, ITHET 2018. 2018; 1–8. https://doi.org/10.1109/ITHET.2018.8424779
  8. 8. Yang XS. Bat algorithm and cuckoo search: A tutorial. Studies in Computational Intelligence. 2013. pp. 421–434.
  9. 9. Rui T, Simon F, Yang XS, Suash D. Wolf search algorithm with ephemeral memory. Seventh International Conference on Digital Information Management (ICDIM 2012). 2012. pp. 165–172.
  10. 10. Karaboga D, Akay B. Artificial Bee Colony (ABC) Algorithm on Training Artificial Neural Networks. 2007 IEEE 15th Signal Process Commun Appl. 2007.
  11. 11. Jr IF, Yang X. A Hybrid Bat Algorithm. 2013;80: 1–7.
  12. 12. Al-Betar MA, Khader AT. A harmony search algorithm for university course timetabling. Ann Oper Res. 2012;194: 3–31.
  13. 13. Zamli KZ, Din F, Kendall G, Ahmed BS. An experimental study of hyper-heuristic selection and acceptance mechanism for combinatorial t-way test suite generation. Inf Sci (Ny). 2017;399: 121–153.
  14. 14. Mirjalili S, Mirjalili SM, Lewis A. Grey Wolf Optimizer. Adv Eng Softw. 2014;69: 46–61.
  15. 15. Zolghadr-Asli B, Bozorg-Haddad O, Chu X. Crow search algorithm (CSA). Studies in Computational Intelligence. 2018.
  16. 16. Odili JB, Kahar MNM, Anwar S. African Buffalo Optimization: A Swarm-Intelligence Technique. Procedia Computer Science. 2015.
  17. 17. El-Ashmawi WH. An Improved African Buffalo Optimization Algorithm for Collaborative Team Formation in Social Network. Int J Inf Technol Comput Sci. 2018;10: 16–29. pmid:30052473
  18. 18. Mirjalili S, Lewis A. The Whale Optimization Algorithm. Adv Eng Softw. 2016;95: 51–67.
  19. 19. Dhiman G, Kaur A. STOA: A bio-inspired based optimization algorithm for industrial engineering problems. Eng Appl Artif Intell. 2019;82: 148–174.
  20. 20. Yang X-S. Bat Algorithm: Literature Review and Applications. Int J Bio-Inspired Comput. 2013;5: 10.
  21. 21. Abd Elaziz M, Oliva D, Xiong S. An improved Opposition-Based Sine Cosine Algorithm for global optimization. Expert Syst Appl. 2017;90: 484–500.
  22. 22. Mirjalili S. SCA: A Sine Cosine Algorithm for solving optimization problems. Knowledge-Based Syst. 2016.
  23. 23. Karaboga D, Akay B, Ozturk C. Artificial Bee Colony (ABC) Optimization Algorithm for Training Feed-Forward Neural Networks. Modeling decisions for artificial intelligence. Springer Berlin Heidelberg; 2007. pp. 318–329.
  24. 24. Arora S, Singh S. Butterfly optimization algorithm: a novel approach for global optimization. Soft Comput. 2019.
  25. 25. Mirjalili SZ, Mirjalili S, Saremi S, Faris H, Aljarah I. Grasshopper optimization algorithm for multi-objective optimization problems. Appl Intell. 2018.
  26. 26. Yang XS. Harmony search as a metaheuristic algorithm. Stud Comput Intell. 2009;191: 1–14. _1
  27. 27. Mirjalili S, Gandomi AH, Mirjalili SZ, Saremi S, Faris H, Mirjalili SM. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv Eng Softw. 2017.
  28. 28. Wolpert DH, Macready WG. No free lunch theorems for optimization. IEEE Trans Evol Comput. 1997.
  29. 29. Tawhid MA, Savsani V. Multi-objective sine-cosine algorithm (MO-SCA) for multi-objective engineering design problems. Neural Comput Appl. 2019;31: 915–929. pmid:30921485
  30. 30. Kaur S, Prashar S. Single Area Unit Commitment Problem using Hybrid Sine-Cosine Algorithm. Int J Sci Eng Technol Res. 2017;6: 101–112.
  31. 31. Sahlo AT, Ewees AA, Hemdan AM, Hassanien AE. Training of Feedforward Neural Networks Using Sine-Cosine Algorithm to Improve the prediction of Liver Enzymes on FIsh Farmed on Nano-selenite. Comput Eng Conf (ICENCO), 2016 12th Int Conf. 2009; 35–40.
  32. 32. Hafez AI, Zawbaa HM, Emary E, Hassanien AE. Sine cosine optimization algorithm for feature selection. Proc 2016 Int Symp Innov Intell Syst Appl INISTA 2016. 2016; 1–5.
  33. 33. Meshkat M, Parhizgar M. A novel weighted update position mechanism to improve the performance of sine cosine algorithm. 5th Iran Jt Congr Fuzzy Intell Syst - 16th Conf Fuzzy Syst 14th Conf Intell Syst CFIS 2017. 2017; 166–171. https://doi.org/10.1109/CFIS.2017.8003677
  34. 34. Abd Elfattah M, Abuelenin S, Hassanien AE, Pan JS. Handwritten Arabic manuscript image binarization using sine cosine optimization algorithm. Advances in Intelligent Systems and Computing. 2017.
  35. 35. Banerjee A, Nabi M. Re-entry trajectory optimization for space shuttle using Sine-Cosine Algorithm. Proc 8th Int Conf Recent Adv Sp Technol RAST 2017. 2017; 73–77. https://doi.org/10.1109/RAST.2017.8002982
  36. 36. Bureerat S, Pholdee N. Adaptive sine cosine algorithm integrated with differential evolution for structural damage detection. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 2017. _6
  37. 37. Singh N, Singh SB. A novel hybrid GWO-SCA approach for optimization problems. Eng Sci Technol an Int J. 2017;20: 1586–1601.
  38. 38. Pasandideh SHR, Khalilpourazari S. Sine Cosine Crow Search Algorithm: A powerful hybrid meta heuristic for global optimization. 2018. Available: http://arxiv.org/abs/1801.08485
  39. 39. Issa M, Hassanien AE, Oliva D, Helmi A, Ziedan I, Alzohairy A. ASCA-PSO: Adaptive sine cosine optimization algorithm integrated with particle swarm for pairwise local sequence alignment. Expert Syst Appl. 2018;99: 56–70.
  40. 40. Khalilpourazari S, Khalilpourazary S. SCWOA: an efficient hybrid algorithm for parameter optimization of multi-pass milling process. J Ind Prod Eng. 2018;35: 135–147.
  41. 41. Gupta S, Deep K. A hybrid self-adaptive sine cosine algorithm with opposition based learning. Expert Syst Appl. 2019;119: 210–230. https://doi.org/10.1016/j.eswa.2018.10.050
  42. 42. Rizk-Allah RM. An improved sine–cosine algorithm based on orthogonal parallel information for global optimization. Soft Comput. 2019;23: 7135–7161.
  43. 43. Qu C, Zeng Z, Dai J, Yi Z, He W. A Modified Sine-Cosine Algorithm Based on Neighborhood Search and Greedy Levy Mutation. Comput Intell Neurosci. 2018;2018. pmid:30073023
  44. 44. Zou Q, Li A, He X, Wang X. Optimal Operation of Cascade Hydropower Stations Based on Chaos Cultural Sine Cosine Algorithm. IOP Conf Ser Mater Sci Eng. 2018;366.
  45. 45. Li S, Fang H, Liu X. Parameter optimization of support vector regression based on sine cosine algorithm. Expert Syst Appl. 2018;91: 63–77.
  46. 46. Reddy KS, Panwar LK, Panigrahi B, Kumar R. A New Binary Variant of Sine–Cosine Algorithm: Development and Application to Solve Profit-Based Unit Commitment Problem. Arab J Sci Eng. 2018;43: 4041–4056.
  47. 47. Oliva D, Hinojosa S, Elaziz MA, Ortega-Sánchez N. Context based image segmentation using antlion optimization and sine cosine algorithm. Multimed Tools Appl. 2018;77: 25761–25797.
  48. 48. Majhi SK. An Efficient Feed Foreword Network Model with Sine Cosine Algorithm for Breast Cancer Classification. Int J Syst Dyn Appl. 2018;7: 1–14.
  49. 49. Mohan NRR. Modified sine cosine algorithm based secure IOT data placement technique in cloud computing environment. 2019 Int Conf Recent Adv Energy-efficient Comput Commun. 2019.
  50. 50. Gupta S, Deep K. Improved sine cosine algorithm with crossover scheme for global optimization. Knowledge-Based Syst. 2019;165: 374–406. https://doi.org/10.1016/j.knosys.2018.12.008
  51. 51. Rajesh KS, Dash SS. Load frequency control of autonomous power system using adaptive fuzzy based PID controller optimized on improved sine cosine algorithm. J Ambient Intell Humaniz Comput. 2019;10: 2361–2373.
  52. 52. Guo W yan, Wang Y, Dai F, Xu P. Improved sine cosine algorithm combined with optimal neighborhood and quadratic interpolation strategy. Eng Appl Artif Intell. 2020;94: 103779.
  53. 53. Wang B, Xiang T, Li N, He W, Li W, Hei X. A Symmetric Sine Cosine Algorithm with Adaptive Probability Selection. IEEE Access. 2020;8: 25272–25285.
  54. 54. Chen H, Wang M, Zhao X. A multi-strategy enhanced sine cosine algorithm for global optimization and constrained practical engineering problems. Appl Math Comput. 2020;369: 124872.
  55. 55. Guesmi T, Farah A, Marouani I, Alshammari B, Abdallah HH. Chaotic sine–cosine algorithm for chance‐constrained economic emission dispatch problem including wind energy. IET Renew Power Gener. 2020;14: 1808–1821.
  56. 56. Jahannoush M, Arabi Nowdeh S. Optimal designing and management of a stand-alone hybrid energy system using meta-heuristic improved sine–cosine algorithm for Recreational Center, case study for Iran country. Appl Soft Comput J. 2020;96: 106611.
  57. 57. Feng Z kai, Niu W jing, Liu S, Luo B, Miao S min, Liu K. Multiple hydropower reservoirs operation optimization by adaptive mutation sine cosine algorithm based on neighborhood search and simplex search strategies. J Hydrol. 2020;590: 125223.
  58. 58. Feng Z kai, Liu S, Niu W jing, Li B jian, Wang W chuan, Luo B, et al. A modified sine cosine algorithm for accurate global optimization of numerical functions and multiple hydropower reservoirs operation. Knowledge-Based Syst. 2020;208: 106461.
  59. 59. Sarwagya K, Nayak PK, Ranjan S. Optimal coordination of directional overcurrent relays in complex distribution networks using sine cosine algorithm. Electr Power Syst Res. 2020;187: 106435.
  60. 60. Raut U, Mishra S. A new Pareto multi-objective sine cosine algorithm for performance enhancement of radial distribution network by optimal allocation of distributed generators. Evol Intell. 2020.
  61. 61. Dasgupta K, Roy PK, Mukherjee V. Power flow based hydro-thermal-wind scheduling of hybrid power system using sine cosine algorithm. Electr Power Syst Res. 2020;178: 106018.
  62. 62. Wan Y, Ma A, Zhong Y, Hu X, Zhang L. Multiobjective Hyperspectral Feature Selection Based on Discrete Sine Cosine Algorithm. IEEE Trans Geosci Remote Sens. 2020;58: 3601–3618.
  63. 63. Neggaz N, Ewees AA, Elaziz MA, Mafarja M. Boosting salp swarm algorithm by sine cosine algorithm and disrupt operator for feature selection. Expert Syst Appl. 2020;145: 113103. pmid:33101827
  64. 64. Ewees AA, Abd Elaziz M, Al-Qaness MAA, Khalil HA, Kim S. Improved Artificial Bee Colony Using Sine-Cosine Algorithm for Multi-Level Thresholding Image Segmentation. IEEE Access. 2020;8: 26304–26315.
  65. 65. Belazzoug M, Touahria M, Nouioua F, Brahimi M. An improved sine cosine algorithm to select features for text categorization. J King Saud Univ—Comput Inf Sci. 2020;32: 454–464.
  66. 66. Moghdani R, Elaziz MA, Mohammadi D, Neggaz N. An improved volleyball premier league algorithm based on sine cosine algorithm for global optimization problem. Eng Comput. 2020.
  67. 67. Lawal AI, Kwon S, Hammed OS, Idris MA. Blast-induced ground vibration prediction in granite quarries: An application of gene expression programming, ANFIS, and sine cosine algorithm optimized ANN. Int J Min Sci Technol. 2021;31: 265–277.
  68. 68. Kuo RJ, Lin JY, Nguyen TPQ. An application of sine cosine algorithm-based fuzzy possibilistic c-ordered means algorithm to cluster analysis. Soft Comput. 2021;25: 3469–3484.
  69. 69. Dash DR, Dash PK, Bisoi R. Short term solar power forecasting using hybrid minimum variance expanded RVFLN and Sine-Cosine Levy Flight PSO algorithm. Renew Energy. 2021;174: 513–537.
  70. 70. Hussain K, Neggaz N, Zhu W, Houssein EH. An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst Appl. 2021;176: 114778.
  71. 71. Mirjalili S. Knowledge-Based Systems SCA: A Sine Cosine Algorithm for solving optimization problems. 2016;96: 120–133.
  72. 72. Hashim FA, Houssein EH, Mabrouk MS, Al-Atabany W, Mirjalili S. Henry gas solubility optimization: A novel physics-based algorithm. Futur Gener Comput Syst. 2019;101: 646–667.