Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Ranking-based hierarchical random mutation in differential evolution

Abstract

In order to improve the performance of differential evolution (DE), this paper proposes a ranking-based hierarchical random mutation in differential evolution (abbreviated as RHRMDE), in which two improvements are presented. First, RHRMDE introduces a hierarchical random mutation mechanism to apply the classic “DE/rand/1” and its variant on the non-inferior and inferior group determined by the fitness value. The non-inferior group employs the traditional mutation operator “DE/rand/1” with global and random characteristics, which increases the global exploration ability and population diversity. The inferior group uses the improved mutation operator “DE/rand/1” with elite and random characteristics, which enhances the local exploitation ability and convergence speed. Second, the control parameter adaptation of RHRMDE not only considers the complexity differences of various problems but also takes individual differences into account. The proposed RHRMDE is compared with five DE variants and five non-DE algorithms on 32 universal benchmark functions, and the results show that the RHRMDE is superior over the compared algorithms.

1 Introduction

In the field of problem optimization, metaheuristic algorithms have been paid more attention and researched by many scholars because of their fast calculation capabilities based on computer simulation and no need for complicated mathematical formula derivation, so they have developed rapidly. In addition to classic Genetic Algorithm (GA) [1], Simulated Annealing (SA) [2], Particle Swarm Optimization (PSO) [3], Differential Evolution (DE) [4], Artificial Bee Colony (ABC) [5], Firefly Algorithm (FA) [6], etc., many new metaheuristic algorithms are also emerging, such as Grey Wolf Optimizer (GWO) [7], Whale Optimization Algorithm (WOA) [8], Monkey King Evolution (MKE) [9], Marine Predators Algorithm (MPA) [10], Gaining Sharing Knowledge based Algorithm (GSK) [11] and Equilibrium Optimizer (EO) [12], etc. Differential evolution (DE) is a branch of metaheuristic algorithms based on population for optimization, which performs parallel search and obtains information in the search space through the differences among individuals in the population. Because of its high efficiency, simplicity and easy operation, DE has been successfully applied in many fields [1318]. However, like other metaheuristic algorithms, when solving some complex optimization problems, DE is also prone to premature convergence and fall into a local optimal solution [19, 20]. To overcome these shortcomings and improve the convergence speed and stability of DE, many scholars have carried out a lot of research work, most of which based on control parameters and mutation strategies [21].

First, many researchers focused on the optimization of DE control parameters, such as population size NP, scaling factor F and crossover probability CR. Liu and Lampinen [22] proposed a fuzzy adaptive differential evolution (FADE), in which individuals successfully entered the next generation and their fitness values were used as inputs to a fuzzy logic controller to adjust F and CR. Noman [23] proposed an adaptive differential evolution Algorithm (aDE), which compared the fitness of the offspring individual with the average fitness value of the parent population. If the former was smaller, F and CR would remain unchanged. Otherwise, they would be randomly selected in the interval [0.1, 1.0] and [0.0, 1.0] respectively. Ghosh [24] proposed an improved differential evolution algorithm with fitness-based adaptation of the control parameters (FiADE), in which the adjustment of F and CR was based on the objective function value of individuals in the population. Sarker [25] proposed a DE with dynamic Parameters Selection algorithm (DE-DPS), in which NP, F and CR were dynamically selected from corresponding candidate sets according to their historical performance for adaptive looping. Tanabe and Fukunaga [26] proposed success-history based adaptive differential evolution (SHADE), which saved the control parameters when the evolution is successful in the historical memories MF, and MCR to guide the selection of future control parameter values. Based on SHADE [26], Viktorin [27] adjusted F and CR based on the Euclidean distance between the trial individual and the original individual. Draa [28] proposed a sinusoidal differential evolution (SinDE), the adjustment of F and CR in which was based on the use of a new sinusoidal framework.

Second, some scholars paid much attention to the improvement of DE’s mutation strategy. Qin et al. [29] proposed a self-adaptive DE (SaDE), which adopted four mutation strategies to generate trial vectors. The selection of mutation strategies would be affected by the performance of these four strategies. Zhang and Sanderson [30] proposed an adaptive differential evolution with optional external archive algorithm (JADE), which introduced a new mutation strategy “DE/current−to−pbest/1”. In contrast to the classic mutation strategy “DE/current−to−best/1”, the best individual in the classic was replaced by an individual randomly selected from the top P%. Gong and Cai [31] proposed an improved adaptive differential evolution with crossover rate repairing technique and ranking-based mutation called Rcr-IJADE, which was an enhanced version of JADE [30]. Epitropakis [32] proposed a proximity-based mutation algorithmic scheme for DE/rand/1. Mallipeddi [33] proposed a differential evolution algorithm with ensemble of parameters and mutation strategies (EPSDE). During the entire evolution process, different mutation strategy pools coexisted with the value pools of each control parameter and competed with each other to produce offspring. Fan and Yan [34] presented an adaptive differential evolution algorithm with discrete mutation control parameters (DMPSADE), which assigned mutation strategies to individuals based on the roulette wheel. Tang [35] proposed a differential evolution with an individual-dependent mechanism (IDE), which contained four mutation operators used in different stages. Wu [36] proposed a differential evolution with multi-population based ensemble of mutation strategies (MPEDE), which partitioned the population dynamics into three indicator subpopulations and one reward subpopulation. Each indicator subpopulation had its own mutation strategy, and the reward subpopulation was applied to the mutation strategy that had performed best so far. Duan [37] proposed a differential evolution algorithm with dual preferred learning mutation (DPLDE), which selected the random mutation vector based on the fitness value and the Euclidean distance between individuals, and then adjusted the mutation strategy according to the randomly selected individual, the optimal individual and the target individual. Mohamed AW and Mohamed AK [38] proposed an adaptive guided differential evolution algorithm with novel mutation for numerical optimization (AGDE), in which individuals were randomly selected from the best 100P% and the worst 100P% of the population and their difference vectors were used as the disturbance in the mutation operation. Besides, the base vector of the mutation operator was randomly selected from the remaining [NP−2(100P%)] individuals. Wei [39] proposed a random perturbation modified differential evolution algorithm (PRMDE), which contained a new "DE/M_pBest−best/1" mutation operator, and the population mutation operation would be implemented based on this proposed mutation operator or the classic mutation operator "DE/best/1". Wang [40] proposed a self-adaptive mutation differential evolution algorithm based on particle swarm optimization (DEPSO), in which a selection probability determined whether an individual would apply “DE/e−rand/1” mutation strategy or PSO mutation strategy.

In order to overcome the problem of premature convergence on DE and improve its performance, a differential evolution with ranking-based hierarchical random mutation in (RHRMDE) is proposed in this paper. Firstly, a hierarchical random mutation mechanism is designed. The upper layer of the model corresponds to a random global mutation strategy, which is applied by superior (NPNWP) individuals in the population sorted by fitness values. NP and NWP represent the population size and the inferior group size, respectively. At the same time, the lower layer of the model designs a random elite mutation strategy, using individuals randomly selected from superior NWP individuals to guide the evolution of NWP inferiors. Secondly, in the adjustment of the control parameters, the complexity differences of diverse problems and individual differences are considered, as well as the evolutionary status of individuals in the previous generation. In addition, the sensitivity of the inferior group size NWP has been analyzed in this paper. Furthermore, the performance of the proposed RHRMDE is verified on 32 benchmark functions by comparing with five DE variants and five non-DE algorithms, and the experiment results demonstrate that RHRMDE outperforms the compared algorithms.

The rest of this paper is structured as follows. Section 2 gives the classic DE algorithm briefly. Section 3 introduces details of the proposed RHRMDE algorithm. The experimental results compared with five DE variants and five non-DE algorithms are reported and analyzed in Section 4. Section 5 draws the conclusions of this work and offers the future research direction.

2 The classic differential evolution algorithm

In DE, the population at generation G consists of NP individuals , i = 1,2,⋯,NP, all of which has D dimensions. After initialization, the entire population generates the corresponding offspring through loop steps of mutation, crossover, selection until certain termination conditions are met.

2.1 Initialization

The jth component of the ith individual in the initial population can be expressed as: (1) where and are the upper and lower bounds of the jth dimension of the ith individual. randj returns a random number distributed between 0 and 1, j indicates that each component of the individual generates a new random number.

2.2 Mutation

After initialization, for each individual , a mutation individual is generated through mutation operation. The classic mutation operator “DE/rand/1” can be expressed as: (2) where F is a scaling factor, r1, r2, r3 are randomly chosen indices within [1, NP] and are different from the current index i, that is, r1≠r2≠r3≠i.

2.3 Crossover

Next, for the target individual and its associated mutation individual , a trial individual is produced by using binomial crossover. (3) where CR(∈[0,1]) determines the contribution of the mutation individual to the trial individual. jrand denotes a random integer in the range of [1, D], which guarantees that at least one component (jrand) in the trial individual comes from the mutation individual.

2.4 Selection

After that, the selection step is executed: between the trial individual and the target individual, the one with the smaller fitness value will be accepted. In other words, if , the trial individual will become the offspring individual , otherwise the target individual will be retained.

(4)

3 RHRMDE algorithm

3.1 Hierarchical mutation mechanism

In RHRMDE, all individuals in the current population are sorted in descending order according to their fitness value, resulting in the new population as follows: , where is with the maximum fitness value; is with the minimum fitness value. Then the ranked population PG is divided into two groups: the non-inferior group composed of the former (NPNWP) individuals and the inferior group composed of the latter NWP. On this basis, a two-layer random mutation mechanism is adopted.

3.1.1 The random global mutation strategy.

Due to its random characteristic, the classic “DE/rand/1” has stronger global search capability and is applied to the non-inferior group at the individual level: (5) where i = {1,2,⋯,(NPNWP)}, r1,r2,r3 differ from each other and range in the set {1,2,⋯,NP}. However, just because of the global and random properties, the classic mutation operator does not make full use of the information contained in superior individuals, which is beneficial for population evolution and convergence [30, 31, 3841]. Then an improved version of the classic mutation strategy is proposed and applied to the “elite” level.

3.1.2 The random elite mutation strategy.

In this strategy, elite individuals guide the evolution of the inferior group. Meanwhile, for the purpose of accelerating convergence, a weighting factor is presented: (6) where i = {(NPNWP+1),⋯,NP}, Wi(∈[0,1]) is the weighting factor, and are elite individuals randomly selected from the former NWP individuals . That is to say, rp1,rp2,rp3 range in the set {1,2,⋯,NWP} and rp1≠rp2≠rp3. In section 3.2, Fi and Wi will be discussed in more detail.

3.2 Control parameter adaptation

In the classic DE, control parameters are the same for all individuals, which have been applied in the individual level by many researchers later. Unlike most studies, the adaptation of control parameters (F, CR, W) in RHRMDE takes into account the complexity differences of various problems and individual differences.

3.2.1 The scaling factor F adaptation.

Due to control the size of the searching area around the base individual, the scaling factor F is one parameter used to balance the tradeoff between global and local search in RHRMDE. The scaling factor F is adjusted for the following purposes. Firstly, since the elite individuals in RHRMDE guide the evolution direction of the entire population, if a better individual can have more global search generations in the solution space, it is more likely to find a better solution, thereby reducing the probability of the population falling into local optima, so the index i representing individual differences is added to the adjustment of F. Secondly, for a simple optimization problem, it may take only a few dozen evolutionary generations to come up with a solution. However, for a complex optimization problem, it may not get close to the correct solution after hundreds of generations, but has already converged and fallen into the local optimal. Therefore, a factor related to the maximum generation number should be considered when calculating F, so that the change curves of F under different maximum generation numbers will not differ too much. Thirdly, using the evolutionary state of the previous generation to intervene in the adjustment of F helps to get rid of local optimality. Therefore, if evolution of the ith individual in the previous generation fails, the Fi of the ith individual in the current ranked population will be assigned a random value. (7) where G and Gmax denote the current and the maximum generation number respectively. is a binary variable that records whether the ith individual in the previous generation has evolved successfully. If the evolution is successful, then ; otherwise, randi denotes a uniform random variable in the interval [0,1]. The random setting of Fi can improve the exploration ability by increasing population diversity.

Without considering the evolutionary status (set all ), the maximum generation number is designed as 100 and 1000 respectively. Taking the 1st, 10th, 20th individuals as examples, the dynamic adaptation curves of the scaling factor F are shown in Fig 1.

thumbnail
Fig 1. The adaptation curves of the scaling factor with different maximum generation numbers and different individuals.

The horizontal axis and the vertical axis are generations and the scaling factor. (a) Gmax = 100,i = 1; (b) Gmax = 100,i = 10; (c) Gmax = 100,i = 20; (d) Gmax = 1000,i = 1; (e) Gmax = 1000,i = 10; (f) Gmax = 1000,i = 20.

https://doi.org/10.1371/journal.pone.0245887.g001

As F approaches zero, the search is conducted over a minimal area around the base vector. Conversely, the larger F is, the larger the search area around the base vector is. Observing Fig 1A–1C and Fig 1D–1F from the horizontal direction, it can be concluded that, with the same maximum generation number, the better the individual (the smaller the index number), the lager the corresponding generation number when F is close to zero. In other words, the better individual has more global search opportunities. By longitudinal comparison of Fig 1A–1C and Fig 1D–1F, it is clear that, although the maximum generation number is different, the dynamic adaptation curves of scaling factors of individuals with the same index are similar. In other words, for optimization problems with different complexity, the ratios of global and local search are similar. These are precisely in line with our design concepts of F.

3.2.2 The crossover probability CR adaptation.

The crossover probability CR indicates the probability of inheriting elements from the mutation individual Vi in the process of constructing the trial individual Ui. Generally speaking, the superior individual is closer to the global optimal than the inferior individual [35]. With a smaller CR, the offspring individual can obtain more useful information by inheriting more elements from a superior parent individual. Instead, a larger CR can generate a more promising offspring individual by accepting more elements from the mutation individual if the parent individual is far away from the optimum. Based on this analysis, CR is calculated as follow: (8) where CRmin is the minimum CR, which is set as 0.1 in this paper. i is the ranked index of . NP denotes the population size, and randi denotes a random number uniformly distributed in the interval [0,1].

3.2.3 The weighting factor W adaptation.

In RHRMDE, looking forward to more local search ability to accelerate convergence, the random elite strategy is proposed with a weighting factor, which is calculated by: (9) where fmin, fmax, fi are the fitness value of the best, worst and ith individuals in the ranked population. In particular, when fmax is equal to fmin, Wi is set to a uniformly distributed random variable within [0,1] to increase diversity. With the deepening of generation, the weighting factor Wi decreases, so that the random elite strategy has a stronger local optimization ability.

The pseudo-code for RHRMDE is given in Fig 2.

4 Experiments and results analysis

4.1 Benchmark test functions

To test the efficiency of RHRMDE, 32 benchmark test functions from literature [4244] are utilized. All the test functions are treated as black-box problems and tested on 30D,100D, the search domain and the global minimum f(*) of these benchmarks can be seen in Table 1. These test functions have many properties, such as unimodality, multimodality, non-separation, separation, shifted, rotated, scalable and non-scalable [21]. In these 32 benchmarks, f1f12 are unimodal functions, f13 is the step function, f14 is the noise quartic function, and f15f32 are multimodal functions.

4.2 Computational complexity

The computational complexity of the RHRMDE is mainly composed of the following parts: initialization, fitness evaluation, mutation, crossover and selection, which are represented by Big-O notation respectively as follows: O(NPD), O(NPGmax), O(NPDGmax), O(NPDGmax), O(NPGmax) where NP is the population size, D denotes the dimension of the specific problem, and Gmax indicates the maximum generation number. In addition, the computational complexity of the fitness ranking and individual position updating before each generation of RHRMDE is O(NPGmax)+O(NPDGmax). Therefore, the RHRMDE has a computational complexity of O(NP∙(3∙DGmax+3∙Gmax+D)), which is simply O(NPDGmax).

4.3 Sensitivity analysis to the inferior group size NWP

The impact of the inferior group size NWP on the performance of RHRMDE is investigated by setting NWP = λNP, and the candidate values for λ are 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, respectively. The Friedman test and Wilcoxon’s rank-sum test [45] are conducted for optimization results under different λ values, and the test results are shown in Fig 3 and Table 2, respectively.

thumbnail
Fig 3. The results of Friedman test with 30 variables over 30 independent runs.

https://doi.org/10.1371/journal.pone.0245887.g003

thumbnail
Table 2. The results of Wilcoxon’s rank-sum test over 30 independent runs.

https://doi.org/10.1371/journal.pone.0245887.t002

From Fig 3, it can be observed that the performance of RHRMDE is best at λ = 0.1. From Table 2, it can be seen that RHRMDE is not sensitive to NWP. Based on these statistical results, λ = 0.1 is considered a more appropriate value. Therefore, NWP = 0.1∙NP is employed in the following series of experiments.

4.4 Parameter settings and involved algorithms

The performance of RHRMDE is compared with five DE-based algorithms like SaDE [29], RcrIJADE [31], MPEDE [36], DPLDE [37], DEPSO [40] and five non-DE-based algorithms like GWO [7], WOA [8], MKE_v3 [9], MPA [10] and EO [12].

For a fair comparison, the maximum generation number Gmax = 1000 is used as the stopping criterion for all algorithms, the population size is set NP = 100, and each function runs independently for 30 times. The other parameters of the comparative algorithms are consistent with their original literature.

4.5 Results analysis

4.5.1 Convergence accuracy analysis.

With the same test function, the average of function error value f(Xbest)−f(X*) represents the convergence accuracy of the algorithm, where Xbest is the best solution found by the algorithm in each run and X* is the real global optimal solution of the test function. Moreover, the Wilcoxon signed-rank test at the 0.05 significance level, the Wilcoxon’s rank-sum test, the Friedman test and Kruskal Wallis test are conducted on the experimental results to obtain a reliable statistic conclusion.

From Tables 3 and 4, the following conclusions can be drawn. (1) RHRMDE obtains global optimal solutions for 30 and 100 dimensional problems of 12 unimodal functions (f1~f12) and 12 multimodal functions (f16~f24, f27~f28, f30). (2) When D = 30, for step function f13, Rcr-IJADE gets the global optimal solution while RHRMDE performs better than DPLDE, DEPSO, GWO, WOA, MKE_v3 and MPA; for quartic noise function f14, EO is the best while RHRMDE performs better than SaDE, MPEDE, DPLDE, DEPSO and MKE_v3; for f15, f25 and f32, Rcr-IJADE is the best while RHRMDE performs better than DPLDE, DEPSO and GWO. In addition, RHRMDE also performs better than SaDE on f15 and better than WOA, MKE_v3 on f32; for f29 and f31, MPEDE is the best while RHRMDE outperforms SaDE, DPLDE, DEPSO, GWO, WOA, EO on f29 and outperforms DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA on f31. (3) When D = 100, for f13 and f32, Rcr-IJADE is the best while RHRMDE outperforms SaDE, DPLDE, DEPSO, GWO,WOA, MKE_v3, MPA, EO on f13 and outperforms DPLDE, DEPSO, GWO on f32; for f14 and f15, EO is the best, RHRMDE is better than SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, MKE_v3 on f14, and slightly worse than MPA, EO on f15; for f25 and f26, WOA is the best while RHRMDE outperforms DPLDE, DEPSO on f25 and outperforms SaDE, Rcr-IJADE, MPEDE; GWO, MKE_v3 on f26; for f29, MPEDE is the best while RHRMDE performs better than SaDE, DPLDE, DEPSO, GWO, MKE_v3 and EO; for f31, RHRMDE obtains the best solution. (4) Contrary to the performance degradation of other algorithms, RHRMDE performs stably and one more optimal solution obtained on high-dimensional problem compared with low-dimensional problem.

thumbnail
Table 3. Mean and STD obtained by SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA, EO and RHRMDE on benchmark functions at 30D.

https://doi.org/10.1371/journal.pone.0245887.t003

thumbnail
Table 4. Mean and STD obtained by SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA, EO and RHRMDE on benchmark functions at 100D.

https://doi.org/10.1371/journal.pone.0245887.t004

As shown in Fig 4, RHRMDE performs better than SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA and EO in 25, 22, 24, 20, 27, 26, 21, 27, 19, 17 out of 32 functions at D = 30, while 30, 28, 28, 19, 25, 28, 21, 30,18, 19 out of 32 functions at D = 100. In contrast, RHRMDE is only worse than SaDE, Rcr-IJADE, MPEDE, DEPSO, GWO, WOA and MKE_v3 in 4, 8, 7, 0, 0, 2, 4, 4, 5, 6 out of 32 functions at D = 30, while 2, 4, 4, 0, 1, 2, 5, 2, 6, 5 out of 32 functions at D = 100. As a result, RHRMDE is generally superior to SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, MKE_v3, MPA and EO on these test functions.

thumbnail
Fig 4. Wilcoxon signed-rank test results at the 0.05 significance level.

(a) D = 30; (b) D = 100.

https://doi.org/10.1371/journal.pone.0245887.g004

According to Table 5, whether D equals 30 or 100, it can be observed that RHRMDE gets higher R+ values than R values for the ten competitors. As can be seen from the computed pvalues, RHRMDE shows a significant improvement in terms of different criteria over SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA and EO at the significance level α = 0.05,01. Thus, it can be concluded that RHRMDE is significantly better than its competitors.

thumbnail
Table 5. The results of Wilcoxon’s rank-sum test compared with ten algorithms over independent 30 runs.

https://doi.org/10.1371/journal.pone.0245887.t005

As can be seen from Fig 5, when D = 30,100, both the Friedman test and the Kruskal-Wallis test show that RHRMDE is the best and MKE_v3 is the worst. In addition, EO performs the second best at D = 30,100 under the Friedman test, while DPLDE performs the second best at D = 30,100 under the Kruskal-Wallis test.

thumbnail
Fig 5. The results of the Friedman test and Kruskal-Wallis test.

(a) Friedman test results; (b) Kruskal-Wallis test results.

https://doi.org/10.1371/journal.pone.0245887.g005

4.5.2 Stability analysis

Due to the stochastic nature of these metaheuristic algorithms, the stability of the proposed algorithm has been analyzed by solution error results over 30 independent runs at D = 30. Box plots of solution error results of all compared algorithms on two unimodal and four multimodal functions are shown in Fig 6. It can be seen that RHRMDE has more concentrated solution error values and fewer outliers compared with other algorithms. Therefore, it can be concluded that RHRMDE is superior to other competitors in terms of stability.

thumbnail
Fig 6. Box plots of the solution error results at D = 30 over independent 30 runs.

(a) f1; (b) f10; (c) f17; (d) f20; (e) f21; (f) f28.

https://doi.org/10.1371/journal.pone.0245887.g006

4.5.3 Convergence speed analysis

In order to compare the convergence speed intuitively, the convergence curves of the mean function error values of all compared algorithms on two unimodal and four multimodal functions at D = 30 are plotted in Fig 7. The difference between performance is pronounced in these convergence curves. The steep slope of the RHRMDE curves indicates that RHRMDE converges rapidly and rarely stops before finding the global optimal solution. Due to the introduction of the random global mutation strategy and control parameter adaptation, the population diversity is preserved in the early generations. Therefore, although the convergence speed of RHRMDE is not the fastest in the beginning, it produces the best results in the end. From the perspective of steepness and the final solution, it can be concluded that the convergence speed of RHRMDE is much faster than that of SaDE, Rcr-IJADE, MPEDE, DPLDE, DEPSO, GWO, WOA, MKE_v3, MPA and EO.

thumbnail
Fig 7. Convergence curves of the mean function error values for six test functions at D = 30.

The horizontal axis and the vertical axis are generations and the mean function error values over 30 independent runs. a) f1. (b) f10. (c) f17. (d) f20. (e) f21. (f) f28.

https://doi.org/10.1371/journal.pone.0245887.g007

4.6 Efficiency analysis of control parameter adaptation

The efficiency analysis for control parameter adaptation in RHRMDE is carried out, while the efficiency of hierarchical mutation strategy can be verified by sensitivity analysis to the inferior group size NWP as described in section 4.3. Some variants of RHRMDE are listed as follows.

  1. To verify the effectiveness of scaling factor F adaptation, RHRMDE variants adopt adaptive CR, W and fixed F = 0.3, F = 0.7 and rand (a uniformly distributed random number in [0, 1]), which are respectively named RHRMDE-1, RHRMDE-2 and RHRMDE-3.
  2. To investigate the effectiveness of crossover probability CR adaptation, RHRMDE variants with adaptive F, W and unaltered CR = 0.3, 0.7, rand are respectively named RHRMDE-4, RHRMDE-5 and RHRMDE-6.
  3. To study the effectiveness of weighting factor W adaptation, RHRMDE variants with adaptive F, CR and settled W = 0.3, 0.7, rand are respectively named RHRMDE-7, RHRMDE-8 and RHRMDE-9.

The performance of these different RHRMDE variants is compared in terms of the Friedman test and Wilcoxon’s rank-sum test, the results of which can be seen in Fig 8 and Table 6, respectively. From Fig 8, it can be observed that the proposed RHRMDE, RHRMDE-6, RHRMDE-5 are the best, second best and third best respectively. From Table 6, the proposed RHRMDE outperforms other RHRMDE variants with fixed F, W and a smaller crossover probability except for RHRMDE-5 with a larger CR and RHRMDE-6 with a random CR. The above experimental comparisons prove that the control parameter adaptation of RHRMDE is effective and its comprehensive effect is the best. It is worth noting that the contribution of the adaptation of the scaling factor and weighting factor is larger than that of the crossover probability.

thumbnail
Fig 8. The Friedman test results of proposed RHRMDE and 9 RHRMDE variants with 30 variables over 30 independent runs.

https://doi.org/10.1371/journal.pone.0245887.g008

thumbnail
Table 6. The Wilcoxon’s rank-sum test results of proposed RHRMDE and 9 RHRMDE variants over 30 independent runs.

https://doi.org/10.1371/journal.pone.0245887.t006

4.7 Discussion on the comparison results

A series of experiments and comparisons have verified the superior performance of RHRMDE. The reasons are summarized as follows: (1) In the hierarchical mutation mechanism, the random global mutation strategy is conducive to extensive searching in the solution space and maintaining population diversity; the random elite mutation strategy makes full use of randomly selected elite individuals to guide the evolution of inferior individuals, thereby positively affecting the evolution direction of the entire population and accelerating the convergence speed. (2) The scaling factor adaptation considers the complexity differences of various problems and individual differences, and adjusts in combination with the historical evolutionary state. Therefore, the proposed algorithm has different global exploration and local exploitation capabilities for each individual and problem, and can better adapt to various optimization environments. (3) The crossover probability adaptation also takes into account individual differences. When the individual is superior, the crossover probability is more likely to be smaller, so there is a greater likelihood of keeping useful information of the superior individual. (4) Under the influence of fitness value and generations, the weighting factor adaptation improves the local search ability of the proposed algorithm and accelerates the convergence speed of the proposed algorithm.

5 Conclusions

A new DE variant algorithm, RHRMDE, is proposed in this paper. In the RHRMDE, the novel hierarchical mutation mechanism based on the non-inferior group and the inferior group is designed to maintain the population diversity and accelerate convergence speed. At the same time, the control parameter adaptation includes historical evolutionary status and individual differences information, considering the complexity differences of various problems, and combining with the hierarchical mutation mechanism to further optimize the algorithm’s balancing ability for exploration and exploitation and speed up the convergence. In addition, the performance of RHRMDE is verified by comparing with five advanced DE variants like SaDE [29], RcrIJADE [31], MPEDE [36], DPLDE [37], DEPSO [40] and five metaheuristic algorithms like GWO [7], WOA [8], MKE_v3 [9], MPA [10] and EO [12] on 32 universal benchmark functions. The experiment results show that: (1) RHRMDE is not sensitive to the inferior group size. (2) In terms of convergence accuracy, stability and convergence speed, RHRMDE performs best for all the unimodal functions and most of the multimodal functions. For the step function, quartic noise function and some multimodal functions, RHRMDE is not the best, which is in line with “No free lunch theorems for optimization” [46]. (3) The adaptive adjustment of the control parameters F, CR and W of RHRMDE is effective, and the integration effect is the best. (4) RHRMDE has the best performance among all the compared algorithms.

As a continuation of this research, the binary and multi-objective version of RHRMDE can be developed. Also, it can be applied to some practical engineering issues, such as airspace planning, flight sequencing, etc.

References

  1. 1. Holland JH. Adaptation in Natural and Artificial Systems. University of Michigan Press; 1975.
  2. 2. Kirkpatrick S, Gelatt CD, Vecchi MP. Optimization by simulated annealing. Science. 1983; 220:671–680. pmid:17813860
  3. 3. Kennedy J, Eberhart RC. Particle swarm optimization. Proceedings of 1995 IEEE international conference on neural networks. 1995; 1942–1948.
  4. 4. Storn R, Price K. Differential evolution: A simple and efficient adaptive scheme for global optimization over continuous spaces. Journal of Global Optimization, 1997, 11(4): 341–359.
  5. 5. Karaboga D, Basturk B. A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm. Journal of Global Optimization. 2007; 39(3):459–71.
  6. 6. Yang XS. Firefly algorithms for multimodal optimization. Stochastic Algorithms: Foundations and Applications, Proceedings. 2009; 5792. 169–178.
  7. 7. Mirjalili S, Mirjalili SM, Lewis A. Grey Wolf Optimizer. Advances in Engineering Software. 2014; 69:46–61.
  8. 8. Mirjalili S, Lewis A. The Whale Optimization Algorithm. Advances in Engineering Software. 2016; 91:51–67.
  9. 9. Meng ZY, Pan JS. Monkey King Evolution: A new memetic evolutionary algorithm and its application in vehicle fuel consumption optimization. Knowledge-Based Systems. 2016; 97:144–157.
  10. 10. Faramarzi A, Heidarinejad M, Mirjalili S, Gandomi AH. Marine Predators Algorithm: A Nature-inspired Metaheuristic. Expert Systems with Applications. 2020; 152:113377.
  11. 11. Mohamed AW, Hadi AA, Mohamed AK. Gaining-sharing knowledge based algorithm for solving optimization problems: a novel nature-inspired algorithm. International Journal of Machine Learning and Cybernetics. 2020; 11:1501–1529.
  12. 12. Faramarzi A, Heidarinejad M, Stephens B, Mirjalili S. Equilibrium optimizer: A novel optimization algorithm. Knowledge-Based Systems. 2020. 191:105190.
  13. 13. Piotrowski AP. Differential Evolution algorithms applied to Neural Network training suffer from stagnation. Applied Soft Computing. 2014; 21: 382–406.
  14. 14. Zamuda A, Sosa JDH, Adler L. Constrained differential evolution optimization for underwater glider path planning in sub-mesoscale eddy sampling. Applied Soft Computing. 2016; 42: 93–118.
  15. 15. Wang Y, Liu H, Long H, Zhang ZJ, Yang SX. Differential Evolution With a New Encoding Mechanism for Optimizing Wind Farm Layout. IEEE Transactions on Industrial Informatics. 2018; 14 (3):1040–1054.
  16. 16. Aguitoni MC, Pavao LP, Ravagnani MADSS. Heat exchanger network synthesis combining Simulated Annealing and Differential Evolution. Energy. 2019; 181: 654–664.
  17. 17. Li SJ, Gong WY, Yan XS, Hu CY, Bai DY, Wang L. Parameter estimation of photovoltaic models with memetic adaptive differential evolution. Solar Energy. 2019; 190:465–474.
  18. 18. Li H, Wang H, Wang L, Zhou XT. A modified Boltzmann Annealing Differential Evolution algorithm for inversion of directional resistivity logging-while-drilling measurements. Journal of Petroleum Science and Engineering. 2020; 188:106916.
  19. 19. Storn R. On the usage of differential evolution for function optimization. Proceedings of the North American Fuzzy Information Process. 1996; 519–523.
  20. 20. Lampinen J, Zelinka I. On stagnation of the differential evolution algorithm. Proceedings of the MENDEL, Brno, Czech, 2000:76–83.
  21. 21. Bilal Millie Pant, Zaheer Hira, Laura Garcia-Hernandez Ajith Abraham. Differential Evolution: A review of more than two decades of research. Engineering Applications of Artificial Intelligence. 2020; 90:103479.
  22. 22. Liu JH, Lampinen J. A Fuzzy Adaptive Differential Evolution Algorithm. Soft Computing. 2005; 9 (6):448–462.
  23. 23. Noman N, Bollegala D, Iba H. An adaptive differential evolution algorithm. 2011 IEEE Congress on Evolutionary Computation. 2011;2229–2236.
  24. 24. Ghosh A, Das S, Chowdhury A, Gini R. An improved differential evolution algorithm with fitness-based adaptation of the control parameters. Information Sciences. 2011;181(18):3749–3765.
  25. 25. Sarker RA, Elsayed SM, Ray T. Differential Evolution With Dynamic Parameters Selection for Optimization Problems. IEEE Transactions on Evolutionary Computation. 2014; 18 (5):689–707.
  26. 26. Tanabe R, Fukunaga A. Success-history based parameter adaptation for differential evolution. 2013 IEEE Congress on Evolutionary Computation. 2013; 71–78.
  27. 27. Viktorin A, Senkerik R, Pluhacek M, Kadavy T, Zamuda A. Distance based parameter adaptation for Success-History based Differential Evolution. Swarm and Evolutionary Computation. 2019; 50:100462.
  28. 28. Draa A, Bouzoubia S, Boukhalfa I. A Sinusoidal Differential Evolution Algorithm for Numerical Optimization. Applied Soft Computing. 2015; 27:99–126.
  29. 29. Qin AK, Huang VL, Suganthan PN. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Transactions on Evolutionary Computation. 2009; 13(2): 398–417.
  30. 30. Zhang J, Sanderson AC. JADE: Adaptive differential evolution with optional external archive. IEEE Transactions on Evolutionary Computation. 2009; 13 (5):945–958.
  31. 31. Gong WY, Cai ZH. Parameter extraction of solar cell models using repaired adaptive differential evolution. Solar Energy. 2013; 94:209–220.
  32. 32. Epitropakis MG, Tasoulis DK, Pavlidis NG, Plagianakos VP, Vrahatis MN. Enhancing Differential Evolution Utilizing Proximity-Based Mutation Operators. IEEE Transactions on Evolutionary Computation. 2011; 15 (1):99–119.
  33. 33. Mallipeddi R, Suganthan PN, Pan QK, Tasgetiren MF. Differential evolution algorithm with ensemble of parameters and mutation strategies. Applied Soft Computing. 2011; 11 (2):1679–1696.
  34. 34. Fan QQ, Yan XF. Self-adaptive differential evolution algorithm with discrete mutation control parameters. Expert Systems with Applications. 2015; 42(3):1551–1572.
  35. 35. Tang LX, Dong Y, Liu JY, Differential evolution with an individual-dependent mechanism. IEEE Transactions on Evolutionary Computation. 2015;19 (4):560–574.
  36. 36. Wu GH, Mallipeddi R, Suganthan PN, Wang R, Chen HK. Differential evolution with multi-population based ensemble of mutation strategies. Information Sciences. 2016; 329:329–345.
  37. 37. Duan MJ, Yang HY, Liu H, Chen JY. A differential evolution algorithm with dual preferred learning mutation. Applied Intelligence. 2019; 49 (2):605–627
  38. 38. Mohamed AW, Mohamed AK. Adaptive guided differential evolution algorithm with novel mutation for numerical optimization. International Journal of Machine Learning and Cybernetics. 2019; 10 (2): 253–277.
  39. 39. Wei ZK, Xie XL, Bao TT, Yu Y. A random perturbation modified differential evolution algorithm for unconstrained optimization problems. Soft Computing. 2019; 23 (15):6307–6321.
  40. 40. Wang SH, Li YZ, Yang HY. Self-adaptive mutation differential evolution algorithm based on particle swarm optimization. Applied Soft Computing. 2019; 81:105496.
  41. 41. Cui LZ, Li GH, Zhu ZX, Lin QZ, Wong KC, Chen JY, et al. Adaptive multiple-elites-guided composite differential evolution algorithm with a shift mechanism. Information Sciences. 2018; 422:122–143.
  42. 42. Suganthan PN, Hansen N, Liang J, Deb Ket al. Problem Definitions and Evaluation Criteria for the CEC 2005 Special Session on Real Parameter Optimization. 2005.
  43. 43. Liang JJ, Qu BY, Suganthan PN, Chen Q. Problem definition and evaluation criteria for the CEC 2015 competition on learning-based real-parameter single objective optimization. 2015.
  44. 44. Awad NH, Ali MZ, Suganthan PN, Liang JJ, Qu BY. Problem definitions and evaluation criteria for the CEC 2017 Special Session and Competition on Single Objective Real-Parameter Numerical Optimization. 2016.
  45. 45. Derrac J, Garcia S, Molina D, Herrera F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evolutionary Computation. 2011; 1(1): 3–18.
  46. 46. Wolpert DH, Macready W. No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation. 1997;1 (1): 67–82.