Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Parameter Estimation, Robust Controller Design and Performance Analysis for an Electric Power Steering System
Previous Article in Journal
Depth Optimization Analysis of Articulated Steering Hinge Position Based on Genetic Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Matrix Adaptation Evolution Strategy with Multi-Objective Optimization for Multimodal Optimization

1
School of Computer Science and Engineering, Xi’an University of Technology, Xi’an 710048, China
2
Shaanxi Key Laboratory for Network Computing and Security Technology, Xi’an 710048, China
Algorithms 2019, 12(3), 56; https://doi.org/10.3390/a12030056
Submission received: 15 January 2019 / Revised: 25 February 2019 / Accepted: 27 February 2019 / Published: 5 March 2019

Abstract

:
The standard covariance matrix adaptation evolution strategy (CMA-ES) is highly effective at locating a single global optimum. However, it shows unsatisfactory performance for solving multimodal optimization problems (MMOPs). In this paper, an improved algorithm based on the MA-ES, which is called the matrix adaptation evolution strategy with multi-objective optimization algorithm, is proposed to solve multimodal optimization problems (MA-ESN-MO). Taking advantage of the multi-objective optimization in maintaining population diversity, MA-ESN-MO transforms an MMOP into a bi-objective optimization problem. The archive is employed to save better solutions for improving the convergence of the algorithm. Moreover, the peaks found by the algorithm can be maintained until the end of the run. Multiple subpopulations are used to explore and exploit in parallel to find multiple optimal solutions for the given problem. Experimental results on CEC2013 test problems show that the covariance matrix adaptation with Niching and the multi-objective optimization algorithm (CMA-NMO), CMA Niching with the Mahalanobis Metric and the multi-objective optimization algorithm (CMA-NMM-MO), and matrix adaptation evolution strategy Niching with the multi-objective optimization algorithm (MA-ESN-MO) have overall better performance compared with the covariance matrix adaptation evolution strategy (CMA-ES), matrix adaptation evolution strategy (MA-ES), CMA Niching (CMA-N), CMA-ES Niching with Mahalanobis Metric (CMA-NMM), and MA-ES-Niching (MA-ESN).

1. Introduction

Many problems from the real world are classified as optimization problems. Some optimization problems that have one global solution are called single modal optimization problems, while others that have multiple global and local optima are known as multimodal optimization problems. Traditional evolutionary algorithms (EAs) are effective at converging to a single global optimum because of the global selection strategy used. However, it is inappropriate for EAs to solve multimodal optimization problems. In order to overcome the weakness, niching techniques are incorporated into EAs, such as differential evolution [1,2], particle swarm optimization [3], the covariance matrix adaptation evolution strategy (CMA-ES) [4], self-adaptive niching CMA-ES [4], and genetic algorithm [5], to solve multimodal optimization problems. The representative niching strategies include crowing [6], restricted tournament selection [7], fitness sharing [8], clearing [9], and speciation [5].
The covariance matrix adaptation evolution strategy (CMA-ES), which was proposed by Hansen and Ostermeier [10], is one of the popular optimization algorithms for solving unconstrained real-parameter optimization. Different from other optimization algorithms, CMA-ES makes use of two evolution paths to realize exploitation and exploration during the search process. The two evolution paths are the learning of the mutation strength and the rank-1 update of the covariance matrix, respectively. The self-adaptively updated covariance matrix, which uses evolution path information, can be considered a time series prediction of the evolution of the parent [11]. Since CMA-ES employs the covariance matrix to exploit the information from the previous and current generations, it has attracted broad investigation in recent years. However, compared with other classical algorithms, such as DE or PSO, CMA-ES is slightly more complicated, because it has two evolution paths and an update of the covariance matrix. To simplify the standard CMA-ES, Beyer and Sendhoff proposed the matrix adaptation evolution strategy (MA-ES) [11], in which one of the evolution paths (namely the p-evolution path) is dropped, and the covariance matrix (namely the C matrix) is discarded. The experimental results in [11] show that the MA-ES exhibits a similar performance as the CMA-ES, which considers both standard population sizes (λ < N) and large population sizes (λ = O (N2)).
The MA-ES only simplifies the CMA-ES. The performance of the MA-ES is basically the same as that of the CMA-ES. Therefore, the MA-ES is a robust local search strategy that efficiently solves unimodal optimization problems. It is unable to find multiple solutions in multimodal problems because of the designed parameters and updating rules [12]. At present, to the best of our knowledge, no work has been reported on utilizing the MA-ES to solve the multimodal problems. There has been an effort to provide two versions of CMA-ES, which are called the niching covariance matrix adaptation evolution strategy and the self-adaptive niching CMA-ES, respectively, for solving multimodal problems [13,14]. The two improved versions of the CMA-ES introduced niching strategies that can maintain the population diversity and realize parallel convergence within some subpopulations to obtain multiple good solutions. However, the performance of the two improved versions of CMA-ES is highly sensitive to niching parameters. So far, some works have been done to convert a multimodal optimization problem (MMOP) to a multi-objective optimization problem (MOP) [15,16,17,18,19]. The advantage of transforming an MMOP into an MOP is that it is unnecessary to use the problem-dependent niching parameters. However, the prerequisite for MOP is objective confliction, which makes it difficult to transform an MMOP into a multi-objective optimization problem [17]. To address this issue, this paper proposes an improved algorithm based on the MA-ES called the matrix adaptation evolution strategy with multi-objective optimization algorithm (MA-ESN-MO). The main contributions of this paper are summarized as follows:
  • The MMOP is transferred into two objective optimization problems with strong objective confliction; the advantage of multi-objective optimization can be fully used to ensure the diversity of the population.
  • The information of the population landscape and the fitness of the objective function are employed to construct two conflicting objective functions instead of utilizing classical niching strategies. Moreover, the archive is employed to save better individuals, which are helpful to ensure the convergence of the algorithm. In this manner, the exploration and exploitation abilities of the algorithm are balanced effectively.
  • The population is divided into several subpopulations in MA-ES instead of using one population in CMA-ES. In this way, the algorithm can explore and exploit in parallel within these subpopulations to find multiple optimal solutions for the given problem. Moreover, the niching method is employed to improve the diversity of the population.
  • Systematic experiments conducted to compare the algorithms including CMA-ES, MA-ES, CMA-ES-Niching (CMA-N), CMA-ES-Niching-MO (CMA-NMO), CMA-ES Niching with Mahalanobis Metric [20] (CMA-NMM), CMA-NMM-MO, MA-ES-Niching (MA-ESN), and MA-ESN-MO on CEC2013 multimodal benchmark problems [21] are described. CMA-NMO and CMA-NMM-MO are obtained by introducing the proposed method into CMA-ES and CMA-NMM. The experimental results show that the proposed method is promising for solving multimodal optimization problems.
The rest of this paper is organized as follows. In Section 2, the related work on multimodal optimization problems is reviewed. Section 3 introduces variants of CMA-ES and the framework of MA-ES. The proposed matrix adaptation evolution strategy with the multi-objective optimization algorithm (MA-ESN-MO) is presented in Section 4. Section 5 reports and discusses the experimental results. Finally, the conclusions and possible future research are drawn up in Section 6.

2. Related Work

Many evolutionary algorithms (EAs) can effectively solve single-objective optimization problems that involve only one optimal solution. However, they are unable to perform well on multimodal optimization problems because of their poor population diversity preservation. To address the issue, sufficient works have been done over the past decades. The strategies on improving the EAs fall into three categories [22].
Niching is an effective method that is used to find and preserve multiple stable niches for multimodal optimization problems. Classical niching techniques include crowding, fitness sharing, speciation, clearing, and restricted tournament selection. The two classic crowding strategies are deterministic crowding and probabilistic crowding. The deterministic crowding can effectively solve the problem of the replacement error, which is the main disadvantage of crowding, while probabilistic crowding utilizes the probabilistic selection to prevent the loss of niches with lower fitness or the loss of local optima [23,24,25]. Both the fitness sharing strategy and speciation divide the population into several subpopulations according to the similarity of the individuals, which can form and maintain the stable niches [5,8]. However, the niche radius σ s h a r e and r s that are used in sharing and speciation, respectively, are difficult to define because they require prior knowledge of the problems. The clearing strategy preserves the best individuals and removes the bad individuals of the niches during the generations [26]. However, the individuals will move toward the best individual area, which may arouse stagnates in the niching because of diversity loss. Restricted tournament selection [7] utilizes Euclidean or Hamming distance to find the nearest member within the w (window size) individuals. The nearest member will compete with the offspring, and the winner will survive in the next generation.
The second strategy aims to enhance population diversity by introducing novel operators into EAs. Among some representative works, Hui et al. [27] proposed an ensemble and arithmetic recombination-based speciation DE (EARSDE) algorithm, where the arithmetic recombination with speciation is used to enhance exploration, and the neighborhood mutation with ensemble strategies is employed to improve the exploitation of individual peaks. Yang et al. [28] proposed an adaptive multimodal continuous ant colony optimization algorithm, where a local search scheme based on Gaussian distribution is used to enhance the exploitation and a differential evolution mutation operator is employed to accelerate convergence. Haghbayan et al. [29] proposed a niche gravitational search algorithm (GSA) method, where a nearest neighbor scheme and the hill valley algorithm are used to enable the species to explore more optima via diversity conservation in the swarm. Qu et al. [30] proposed a distance-based locally informed particle swarm optimization (PSO) algorithm, where several local best particles are used to guide the search of each particle instead of using the global best particle.
The third strategy is to introduce a novel transformation mechanism, in other words, a multimodal optimization problem (MMOP) is transformed into a multi-objective optimization problem (MOP) [17]. Among some representative works, Cheng et al. [22] proposed an evolutionary multi-objective optimization-based multimodal optimization algorithm, where approximate multimodal fitness landscapes are used to provide an estimation of potential optimal areas. Moreover, an adaptive peak detection strategy is employed to find peaks where optimal solutions may exist. Yu et al. [19] proposed a tri-objective differential evolution approach algorithm, where three optimization objectives are constructed to ensure good population diversity. In addition, a solution comparison rule and a ranking strategy are employed to enhance the accuracy of solutions. Wang et al. [17] proposed a multi-objective optimization for multimodal optimization problems (MOMMOP) algorithm, where an MMOP is transformed into a multi-objective optimization problem with two conflicting objectives. Basak et al. [15] proposed a novel multimodal optimization algorithm, where a novel bi-objective formulation of the multimodal optimization problem and differential evolution (DE) with a non-dominated sorting strategy are used to detect multiple global and local optima. Deb et al. [16] proposed a bi-objective evolutionary algorithm, where the single-objective multimodal optimization problem is converted into a suitable bi-objective optimization problem to find multiple peaks. Yao et al. [18] proposed a multipopulation genetic algorithm, where a multipopulation and clustering scheme are used to improve exploitation within promising areas.

3. Variants of CMA-ES and MA-ES Algorithm

This section briefly reviews the improvements on CMA-ES and the framework of the MA-ES algorithm, respectively.

3.1. Variants of CMA-ES

In recent years, many CMA-ES variants have been developed to solve single-objective and multi-objective optimization problems. For single-objective optimization problems, a DE variant with covariance matrix self-adaptation (DECMSA) [31] combines different features of DE and CMA-ES instead of hybridizing these two algorithms simply. In DECMSA, the individuals are sampled from a Gaussian distribution to guide the search direction in the DE. In addition, an enhanced local search strategy based on CMA-ES is designed. In order to improve the classical optimization algorithm’s ability of rotational invariance, differential evolution based on covariance matrix learning and the bimodal distribution parameter setting (CoBiDE) [32] utilizes the covariance matrix learning to construct an appropriate coordinate system for the crossover operator, while biogeography-based optimization (BBO) with covariance matrix-based migration (CMM-BBO) [33] utilizes the covariance matrix migration to reduce BBO’s dependence on the coordinate system. The differential covariance matrix adaptation evolutionary algorithm (DCMA-EA) [34] incorporated the mutation, crossover, and selection operators of the DE into CMA-ES to compose a hybrid algorithm in order to improve the performance of CMA-ES for solving the optimization problems with complicated fitness landscapes. The bilevel covariance matrix adaptation evolution strategy (BL-CMA-ES) [35] employs CMA-ES at the upper level and lower level to extract the priori knowledge: the search distribution, which significantly reduces the number of function evaluations and improves the efficiency of the algorithm. The Differential Crossover Strategy based on Covariance Matrix Learning with Euclidean Neighborhood (L-covnSHADE) [36] employs covariance matrix learning to establish a coordinate system for the better crossover operator.
For multi-objective optimization problems (MOPs), multi-objective differential evolution (MODE) with dynamic covariance matrix learning (MODEs + DCML) [37] utilizes the dynamic covariance matrix learning (DCML) to establish a proper coordinate system for the binomial crossover operator. Therefore, the poor performance of MOPs solved by MODE can be significantly improved. Decomposition-based multi-objective evolutionary algorithms (MOEA/D)-CMA-ES [38] integrates CMA-ES into the decomposition-based multi-objective evolutionary algorithms (MOEA/D) to solve the multi-objective optimization problems. The hybrid algorithm MOEA/D-CMA-ES takes advantage of the MOEA/D in multi-objective optimization and CMA-ES in complex numerical optimization. The experimental results show that MOEA/D–CMAES is an effective algorithm for solving complex multi-objective optimization problems.

3.2. MA-ES Algorithm

The CMA-ES employs two evolution paths: one for the update of the covariance matrix and another for the learning of the covariance matrix. This increases the complexity of the algorithm. To solve this problem, the MA-ES drops one of the evolution paths and removes the covariance matrix. Briefly speaking, the MA-ES directly learns the M matrix from the path cumulation information. Different from CMA-ES, the offspring of the new generation g + 1 is calculated by weighted recombination, which is expressed by the following formula:
y g + 1 = x g + σ g d ˜ g w
where σ g is the global step size or mutation strength at generation g, and d ˜ g is considered as a search direction vector.
The path cumulation s g + 1 can be detailed as follows [11]:
s g + 1 = ( 1 c s ) s g + μ e f f c s ( 2 c s ) z ˜ g w
where z ˜ g w = ( C g ) 1 2 y g + 1 y g σ g . C is the covariance matrix. The detailed derivation of z ˜ g w can be found in [11]. c s can be regarded as memory time constants [12], which is defined as follows [11]:
c s = μ e f f + 2 μ e f f + D + 5
where D is the search space dimension, and μ e f f denotes the variance effective selection mass [12], which is calculated as follows [11]:
μ e f f = 1 k = 1 μ w m 2
where μ = λ 2 , λ = 4 + 3 l n D , w m = 1 μ .
The learning process of the M matrix instead of the covariance matrix C of CMA-ES is updated according to the following equation [11]:
M g + 1 = M g [ I + c 1 2 ( s g + 1 × ( s g + 1 ) T I ) + c w 2 ( z ˜ g × ( z ˜ g ) T w I ) ]
where I is the identity matrix, c w = min ( 1 c 1 ,   α c o v μ e f f + 1 μ e f f 2 ( D + 2 ) 2 + α c o v × μ e f f 2 ) , c 1 = α c o v ( D + 1.3 ) 2 + μ e f f , and α c o v = 2 .
The step-size of the mutation σ g is updated according to the following equation [11]:
σ g + 1 = σ g exp [ c s d σ ( s g + 1 E [ N ( 0 , I ) ] 1 ) ]
where E [ N ( 0 , I ) ] = 2 Γ ( D + 1 2 ) / Γ ( D 2 ) , d σ = 1 + c s + 2 × max ( 0 , μ e f f 1 D + 1 1 ) .
The pseudo-code of the MA-ES algorithm is shown in Algorithm 1 [11].
Algorithm 1. MA-ES algorithm.
1:Initialize (y(0), σ(0), g = 0, s(0) := 0, M(0) = I)
2:while termination condition(s) is not fulfilled
3:  for l = 1 to λ do
4:    z ˜ l ( g ) = N l ( 0 , I )
5:    d ˜ l ( g ) = M ( g ) z ˜ l ( g )
6:    f ˜ l ( g ) = f ( y ( g ) + σ ( g ) d ˜ l ( g ) )
7:  end for
8:  SortOffspringPopulation
9:  Update y(g+1) according to (1)
10:  Update s(g+1) according to (2)
11:  Update M(g+1) according to (5)
12:  Update σ(g+1) according to (6)
13:  g = g + 1
14:end while

4. The MA-ESN-MO Algorithm

Based on the above observations, this section introduces the proposed algorithm MA-ESN-MO. First of all, we introduce the transformation strategy, which transforms an MMOP into an MOP. Then, the MA-ES is employed as a search engine to generate offspring. To obtain solutions with high accuracy, the archive is employed to save better individuals so that the quality of the solutions can be ensured. Finally, the dynamic peak identification strategy is used to enhance the diversity of the population.

4.1. Transforming an MMOP into an MOP

The premise of employing multi-objective techniques is that different objectives should conflict with each other. Therefore, two strongly conflicting objectives are constructed as follows:
{ f 1 ( x ) = α × f ( x ) n o r m + x f 2 ( x ) = α × f ( x ) n o r m x
where α is a scaling factor that gradually increases during evolution, f ( x ) n o r m denotes the normalized objective function values, and x is the decision variable. An MMOP is transformed into D bi-objective optimization problems, where D is the dimension of an MMOP.
In Equation (7), if the value of x increases, the value of x will decrease, and vice versa. Moreover, α and f ( x ) n o r m are positive, which brings the same change for x and x . It can be concluded that f 1 ( x ) conflicts with f 2 ( x ) . Therefore, the multi-objective optimization methods can be used. f 1 ( x ) and f 2 ( x ) are mainly influenced by x and x at the early stage of evolution. Population distribution can influence the diversity of the population. Gradually, f 1 ( x ) and f 2 ( x ) are mainly influenced by the fitness at the later stage of evolution. Therefore, the algorithm can quickly converge to find the multiple optimal solutions. Similar to the mapping relationship between the decision space and objective space in a MOP, an example of the mapping relationship between an MMOP and an MOP is shown in Figure 1 (D = 2). α is designed as follows:
α = D × ( b a ) × ( F E S M a x F E S ) D
where D denotes the dimension of the problem, [a, b]D denotes the range of decision space, and FES and MaxFES denote the number of function evaluations and the maximum number of function evaluations, respectively.
Figure 2 shows a transformation example of the equal maxima function. The equal maxima function has five global optima. Figure 2a shows the distribution of the population (x) and their fitness (f(x)) when FES = 1000. Figure 2b shows the results of the transformation from multimodal optimization to bi-objective optimization. Figure 2c shows the result of the non-dominated sorting procedure, which is employed to find a set of representative Pareto optimal solutions. The representative Pareto optimal solutions will be used as the parents of the next generation. Figure 2d shows the distribution of the individuals and their fitness corresponding to the Pareto optimal solutions.

4.2. Matrix Adaptation Evolution Strategy with Multi-Objective Optimization Algorithm

The exploration efficiency is dependent on the distribution of individuals, namely the population diversity. The exploitation efficiency is associated with the fitness of individuals. Then, diversity and fitness can be used to indirectly measure the exploration and exploitation, respectively. Therefore, diversity and fitness can be employed to achieve a trade-off between exploration and exploitation [39]. In view of this idea, the diversity and the fitness are considered in Equation (7). In the proposed algorithm, the non-dominated sorting mechanism is used to find the non-dominated solutions that are helpful for exploring the solution space efficiently and exhaustively. The non-dominated solutions, which take into account fitness and diversity, can be used as the seeds of niching. The non-dominated sorting procedure comes from the non-dominated sorting genetic algorithm II (NSGA-II) [40], which is a classical multi-objective optimization algorithm. Details of the fast non-dominated sorting algorithm can be found in [40].
In MA-ES, the offspring is yielded by the parent that is mutated according to a normal distribution. Then the parent will be discarded, while the excellent offspring will be preserved as the parent for the next generation. However, sometimes, the offspring may achieve worse performance than their parents. Take f2 equal maxima (1D) as an example; f2 has five global optima. Figure 3 shows three independent runs of MA-ES on f2 (the equal maxima 1D function). It can be seen that the number of global optima is not stable during the evolution. Originally, the parents have found the global optima. However, the parents are not preserved. The offspring are only near the optimal solution. Therefore, the number of the global optima found by the algorithm is variable during the evolution. In order to solve the problem, the archive is employed to save the individuals that have better performance. Specifically speaking, if the offspring performs worse, the individuals in the archive will be used as the parents for the next generation.
Based on the above explanation, the pseudo-code of the MA-ESN-MO algorithm is illustrated in Algorithm 2, from which we can see that the MA-ESN-MO generation cycle is performed within the while-end while-loop (line 3 to line 26). C(g) denotes the number of global optima found at generation g. At each generation, a number of λ offspring is generated according to Equation (1), where the parental stage xg is mutated according to a normal distribution yielding offspring yg. For a given individual i, f(yi) can be calculated from the original objective function of MMOP. Then, the population is sorted according to the individual fitness in line (6). Afterward, the two objective values of the transformed bi-objective optimization problem can be obtained according to Equations (7) and (8). Then, the seeds of niching (also referred to as the parent of the next generation) are generated by the non-dominated sorting procedure or dynamic peak identification algorithm according to the condition in line (8). Then, the path cumulation s, matrix M, and the step size σ (also referred to as mutation strength) are updated according to Equations (2), (5) and (6), respectively. At each generation, the excellent individual will be preserved in the archive as the parent for the next generation (line 18, line 23). Therefore, if C(g) is less than C(g−1), this shows that the performance of offspring is worse than that of their parents. Then, the individuals in the archive will be used as the parents for the next generation (line 21).
Algorithm 2. MA-ESN-MO algorithm.
1:Initialize D(number of dimensions), λ, y(0), σ(0), s(0), d(0), M(0) and NP, g = 0
2:Initialize NP parents
3:while the termination condition is not satisfied
4:  Generate the individuals according to Equation (1)
5:  Calculate the objective function value of each individual
6:  Sort the population according to the objective function value
7:  Compute the f1 and f2 values of each dimension according to equations (7) and (8)
8:  if rand > 0.5
9:    Generate the seeds of niching according to the non-dominated sorting procedure
10:  else
11:    Generate the seeds of niching with the dynamic peak identification algorithm (Algorithm 2)
12:  end if
13:  Update s(g+1) according to Equation (2)
14:  Update M(g+1) according to Equation (5)
15:  Update σ(g+1) according to Equation (6)
16:  g = g + 1
17:  if g = = 1
18:    Archive = x(g)
19:  else
20:    if C(g) < C(g−1)
21:      x(g) = Archive
22:    else
23:      Archive = x(g)
24:    end if
25:  end if
26:end while
Output: the global optima with the maximum objective function value in the population.
The pseudo-code of dynamic peak identification is presented in Algorithm 3. Firstly, the population is sorted according to the objective function value. For a given individual Pop{i}, using an estimated so-called niche radius ρ, Pop{i} is classified into a peak and populates this niche (line 12 to line 15). In addition, every niche includes several individuals (line 7 to line 11). Then, the various fitness peaks are identified dynamically within the for-end for-loop (line 6 to line 16).
Algorithm 3. Dynamic Peak Identification [15].
1:Input niche radius ρ, population Pop, and population size NP
2:Sort Pop according to the objective function value
3:NumPeak = 1
4:DPS = {Pop{1}}
5:Niche (NumPeak) = {Pop{1}}
6:fori = 2 to NP
7:  for k = 1 to NumPeak
8:    if Pop{i} and DPS (k) belong to the same niche
9:      Niche (k) = Niche (k)∪{Pop{i}}
10:    end if
11:  end for
12:  ifPop{i} is not within ρ of peak in DPS
13:    NumPeak = NumPeak + 1
14:    DPS = DPS∪{Pop{i}}
15:  end if
16:end for
Output: DPS and Niche
As mentioned earlier, MA-ESN-MO is proposed by introducing three improvement strategies to solve multimodal optimization problems: a transformation strategy, an archive strategy, and a dynamic peak identification strategy. The transformation strategy is used to transform an MMOP into an MOP. The advantage of the transformation strategy is that it is unnecessary to use the problem-dependent niching parameters. The archive is used to save better individuals, which is effective in preserving the seeds of niching. The dynamic peak identification strategy is employed to prevent the best niche from occupying the population’s resources.
There are three main differences among the MA-ESN-MO and the other algorithms mentioned before [4,15,16,17,18,19].
(1)
Two conflicting objective functions are constructed differently. The landscape information of the population is usually helpful for judging the diversity of the population. The fitness of the objective function is helpful for obtaining the global optimum. Therefore, the two strongly conflicting objectives are designed by the landscape information of the population and the fitness of the objective function. In order to find multiple optimal solutions, exploration should be paid attention to in the early stage of evolution, and exploitation should be considered in the later stage of evolution. Therefore, the parameter α, which is proportional to the number of function evaluations, is introduced to adjust for exploration and exploitation.
(2)
The archive is employed to save the seeds of niching. In the classical CMA-ES algorithm and MA-ES algorithm, the parent will be discarded after the offspring are yielded. The CMA-ES and MA-ES perform well in solving unimodal optimization problems because of their strong exploratory ability. In multimodal optimization problems, the parent is responsible for finding the area where the potential optimal solution existed, while the offspring are responsible for exploitation. However, sometimes the offspring perform worse than their parents. As a result, the offspring may fail to find the optimal solution. As the evolution continues, the area where the potential optimal solution existed may be abandoned. Therefore, it will be difficult for the population to find all the global optimal solutions. To alleviate this issue, the archive is introduced in the MA-ESN-MO to save better individuals. At each generation, the individuals in the archive will be used as the parents for the next generation. If the optimal individual from offspring performs better than its parent, the individual will be saved in the archive. Otherwise, the parent will be saved in the archive. Then, it ensures the population to find all the global optimal solutions.
(3)
The non-dominated solutions include all the multiple optima of an MMOP. However, they also include some inferior solutions. Then, the best niche may occupy the population’s resources during the evolution. To address this issue, the dynamic peak identification strategy is used to avoid converging toward a single global optimum. However, the performance of the dynamic peak identification strategy is highly sensitive to the niching radius. In other words, the performance of the algorithm deteriorates with an inappropriate niching radius. Therefore, the transformation strategy and the dynamic peak identification strategy are dynamically employed in MA-ESN-MO to improve the performance of the algorithm.

5. Experiments and Discussions

To verify the effectiveness of the improvement strategies proposed in this paper, 20 test problems from CEC2013 [21] multimodal benchmarks are used. The algorithms for testing include CMA-ES [4], MA-ES [11], CMA-N [14], CMA-NMO, CMA-NMM [20], CMA-NMM-MO, MA-ESN [11,15], and MA-ESN-MO. CMA-NMO, CMA-NMM-MO, and MA-ESN-MO are produced by introducing the improvement strategies into CMA-ES, CMA-NMM, and MA-ES, respectively. To obtain an unbiased comparison, all the experiments are run on a PC with an Intel Core i7-3770 3.40 GHz CPU and 4 GB memory. All of the experiments are run 25 times, and the codes are implemented in Matlab R2013a (MathWorks, Natick, MA, USA).

5.1. Parameter Settings and Performance Criteria

Table 1 shows a brief description of testing problems. f1f10 are simple and low-dimensional multimodal problems, while f11f20 are composition multimodal problems composed of several basic problems with different characteristics. f1f5 have a small number of global optima, and f6f10 have a large number of global optima. For each algorithm, the subpopulation size is set to 10. Some functions from the CEC2013 are drawn in the following. The equal maxima (f2) has five global optima. There are no local optima, as shown in Figure 4. The Himmelblau function (f4) has four global optima, as shown in Figure 5. Figure 6 shows an example of the Shubert 2D function (f6), where there are 18 global optima in nine pairs. Figure 7 shows the 2D version of Composition Function 2 (CF2). CF2 (f12) is constructed based on eight basic functions (n = 8), thus it has eight global optima.
In this experiment, three criteria [6] are employed to measure the performance of different multimodal optimization algorithms on each function.

5.1.1. Peak Ratio

The peak ratio (PR) is used to measure the average percentage of all known global optima found over NR independent runs:
P R = r u n = 1 N R N G F r u n N G O N R
where NGFrun is the number of global optima found in the run-th run, NGO is the number of true global optima, and NR is the number of runs (NR = 25).

5.1.2. Success Rate

The success rate denotes the percentage of successfully detecting all the global optima out of NR runs for each function.

5.1.3. Average Number of Peaks Found

The average number of peaks (ANP) denotes the average number of peaks found by an algorithm over NR runs.
A N P = r u n = 1 N R N G F r u n N R

5.2. Experimental Results of Eight Optimization Algorithms

The experimental results and analyses are shown in the following. The best solution of every single function in different algorithms is highlighted with boldface. The results of the test functions from f1 to f10 are shown in Table 2. As can be seen, the performance of the CMA-ES is the same as MA-ES on f2f5, f7, and f10. The CMA-ES performs slightly better than the MA-ES on f1, f6, f8, and f9. As mentioned earlier, the MA-ES is a simplified version of the CMA-ES; therefore, their performance is similar. The CMA-NMO performs better than the CMA-N on f2, f4, f5, f6, f8, and f10. The CMA-N beats the CMA-NMO on f7 and f9 when ε = 0.1. The CMA-NMM-MO performs better than the CMA-NMM on f4f10. The CMA-NMM outperforms the CMA-NMM-MO on f2 and f3. The MA-ESN-MO performs better than the MA-ESN on f2f6, f8, and f10. The MA-ESN beats the MA-ESN-MO on f7 and f9 when ε = 0.1. The experimental results of the test functions from f1f10 show that the performance of the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO has been greatly improved because of the introduction of the multi-objective optimization strategy, the dynamic peak identification strategy, and the archive.
Table 3 shows the results of the test function from f11 to f20. As can be seen, the CMA-ES performs slightly better than the MA-ES. This suggests that the two evolution paths are more effective in solving complex problems. The CMA-NMO performs better than the CAM-N on f12, but worse on f20. For functions f11f19, the CMA-NMO performs better than the CMA-N on different ε except for ε = 0.1. For function f11f20, the CMA-NMM-MO performs better than the CMA-NMM except for function f19 and f20. The MA-ESN-MO performs better than the MA-ESN on f12 but worse on f19 and f20 than the MA-ESN. For functions f11, f13f18, the MA-ESN-MO performs better than the MA-ESN on different ε except ε = 0.1. The experimental results show that the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO can find better solutions than the CMA-ES, MA-ES, CMA-N, CMA-NMM, and MA-ESN.
Table 4 and Table 5 shows the results of ANP in terms of the mean value (Mean), Peak ratio (PR), and Success Rate (SR) obtained in 25 independent runs by each algorithm for functions f1 to f10 and f11 to f20, respectively. In view of statistics, the Wilcoxon signed-rank test [41] at the 5% significance level is used to compare the MA-ESN-MO with other compared algorithms. “≈”, “+”, and “–” are applied to express the performance of the MA-ESN-MO as similar to, worse than, and better than that of the compared algorithm, respectively. The statistical results are reported in Table 4 and Table 5. Table 4 shows that all the algorithms can accurately find multiple optimal solutions on f1 except for the MA-ES. The CMA-NMM performs better than the other algorithms on function f2. For function f3, all of the algorithms can accurately find optimal solutions except for CMA-N and MA-ESN. The CMA-NMO, CMA-NMM-MO, and MA-ESN-MO perform better than other algorithms on f4f10. For f1f10, the MA-ESN-MO works better than the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, and MA-ESN on eight, eight, eight, zero, seven, zero, and eight test problems, respectively. Table 5 shows that the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO perform better than other algorithms on f11f17. The CMA-NMO performs the best on f18. For function f19, the performance of all the algorithms is basically the same except for the CMA-ES and MA-ES. None of the algorithms can find multiple optimal solutions on f20. For f11f20, the MA-ESN-MO works better than the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, and MA-ESN on nine, nine, seven, zero, seven, one, and seven test problems, respectively.
In order to test the statistical significance of the eight compared algorithms, the Wilcoxon’s test at the 5% significance level, which is implemented by using KEEL software [42], is employed based on the PR values. Table 6 summarizes the statistical test results. It can be seen from Table 6 that the MA-ESN-MO provides higher R+ values than R− values compared with the CMA-ES, MA-ES, CMA-N, CMA-NMM, and MA-ESN. Furthermore, the p values of the CMA-ES, MA-ES, CMA-N, CMA-NMO, and MA-ESN are less than 0.05, which means that the MA-ESN-MO is significantly better than these competitors. The p values of the CMA-NMO and CMA-NMM-MO are equal to one, which means that the performance of the MA-ESN-MO is not different from that of the CMA-NMO and CMA-NMM-MO. To further determine the ranking of the eight compared algorithms, the Friedman’s test, which is also implemented by using KEEL software, is conducted. As shown in Table 7, the overall ranking sequences for the test problems are the CMA-NMO, CMA-NMM-MO, MA-ESN-MO, CMA-N, CMA-NMM, MA-ESN, CMA-ES, and MA-ES. The experimental results show that the improved algorithms CMA-NMO, CMA-NMM-MO, and MA-ESN-MO perform better than other original algorithms. Therefore, it can be concluded that the improvement strategies are effective. Figure 8 shows the results of ANP obtained in 25 independent runs by each algorithm for functions f1f20 on ε = 0.0001. In order to show these clearly in Figure 8, the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN and MA-ESN-MO are abbreviated to CMA, MA, CN, CO, CMM, CMO, MAN, and MMO, respectively.
Table 8 shows the success rate of all the algorithms in finding all the global optimal solutions from f1 to f10. It can be observed that the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO are able to achieve the success rate of 100% on f1f5. Moreover, the CMA-NMO generates a higher success rate than the CMA-N from f6f10. Similarly, the CMA-NMM-MO and MA-ESN-MO generate higher success rates than the CMA-NMM and MA-ESN from f6f10, respectively. The CMA-ES and MA-ES obtain the success rate of 100% on f1 and f3. However, the CMA-ES and MA-ES generate relatively low success rates on other eight test functions.
Figure 9 displays the average number of peaks in terms of the mean value achieved by each of eight algorithms on ε = 0.0001 for CEC2013 multimodal problems versus the number of FES. It can be seen that the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO perform better than other algorithms, which suggests that the mechanism of multi-objective optimization can help the algorithm find multiple global optimal solutions. In addition, the curves of ANP achieved by CMA-ES, MA-ES, CMA-N, CMA-NMM, and MA-ESN are ups and downs instead of steady growth. The reason is that the better parents are discarded after producing the offspring. However, sometimes, the performance of the parent is better than that of the offspring. The curves of the CMA-NMO, CMA-NMM-MO, and MA-ESN-MO show a gradual upward trend. The reason is that the archive is introduced in these algorithms, which is helpful to ensure the convergence of the algorithm.

6. Conclusions

The CMA-ES has received considerable attention as an efficient optimization algorithm. The MA-ES might be more attractive because of a simpler operator compared to the CMA-ES. Although niching techniques have been introduced into the CMA-ES, the performance of the CMA-ES is unsatisfactory for solving multimodal optimization problems. This paper proposed a matrix adaptation evolution strategy with the multi-objective optimization algorithm to solve multimodal optimization problems. The strategy of the multi-objective optimization is used to ensure the diversity of population. The archive is employed to maintain the peak found by the algorithm until the end of the run. The population is divided into multiple subpopulations, which are used to explore and exploit in parallel to find multiple optimal solutions. The experimental results suggest that the proposed strategies can achieve a better performance than the original algorithm on CEC2013 test problems.

Funding

This research is partly supported by the Doctoral Foundation of Xi’an University of Technology (112-451116017).

Acknowledgments

Thanks to Ofer M. Shir for providing the source code of CMA-N and CMA-NMM.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Singh, G.; Deb, K. Comparison of multimodal optimization algorithms based on evolutionary algorithms. In Proceedings of the Genetic and Evolutionary Computation Conference, Seattle, WA, USA, 8–12 July 2006; pp. 1305–1312. [Google Scholar]
  2. Li, X. Efficient differential evolution using speciation for multimodal function optimization. In Proceedings of the Genetic and Evolutionary Computation Conference, Washington DC, USA, 25–29 June 2005; pp. 873–880. [Google Scholar]
  3. Li, X. Niching without niching parameters: Particle swarm optimization using a ring topology. IEEE Trans. Evol. Comput. 2010, 14, 150–169. [Google Scholar] [CrossRef]
  4. Shir, O.M.; Emmerich, M.; Back, T. Adaptive niche radii and niche shapes approaches for niching with the CMA-ES. Evol. Comput. 2010, 18, 97–126. [Google Scholar] [CrossRef] [PubMed]
  5. Li, J.P.; Balazs, M.E.; Parks, G.T.; Clarkson, P.J. A species conserving genetic algorithm for multimodal function optimization. Evol. Comput. 2002, 10, 207–234. [Google Scholar] [CrossRef] [PubMed]
  6. Thomsen, R. Multimodal optimization using crowding-based differential evolution. In Proceedings of the IEEE Congress on Evolutionary Computation, Portland, OR, USA, 19–23 June 2004; pp. 1382–1389. [Google Scholar]
  7. Harik, G.R. Finding multimodal solutions using restricted tournament selection. In Proceedings of the International Conference on Genetic Algorithms, Pittsburgh, PA, USA, 15–19 June 1995; pp. 24–31. [Google Scholar]
  8. Goldberg, D.E.; Richardson, J. Genetic algorithms with sharing for multimodal function optimization. In Proceedings of the International Conference on Genetic Algorithms, Cambridge, MA, USA, 6–8 July 1987; pp. 41–49. [Google Scholar]
  9. Petrowski, A. A clearing procedure as a niching method for genetic algorithms. In Proceedings of the IEEE Congress on Evolutionary Computation, Nagoya, Japan, 20–22 May 1996; pp. 798–803. [Google Scholar]
  10. Hansen, N.; Müller, S.D.; Koumoutsakos, P. Reducing the Time Complexity of the Derandomized Evolution Strategy with Covariance Matrix Adaptation (CMA-ES). Evol. Comput. 2003, 11, 1–18. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Beyer, H.G.; Sendhoff, B. Simplify Your Covariance Matrix Adaptation Evolution Strategy. IEEE Trans. Evol. Comput. 2017, 21, 746–759. [Google Scholar] [CrossRef]
  12. Hansen, N.; Kern, S. Evaluating the CMA Evolution Strategy on Multimodal Test Functions. In Parallel Problem Solving from Nature—PPSN VIII; Springer: Berlin/Heidelberg, Germany, 2004; pp. 282–291. [Google Scholar]
  13. Shir, O.M.; Bäck, T. Niche Radius Adaptation in the CMA-ES Niching Algorithm. In Parallel Problem Solving from Nature—PPSN IX; Springer: Berlin/Heidelberg, Germany, 2006; pp. 142–151. [Google Scholar]
  14. Shir, O.M.; Back, T. Dynamic niching in evolution strategies with covariance matrix adaptation. In Proceedings of the IEEE Congress on Evolutionary Computation, Edinburgh, UK, 2–5 September 2005; Volume 3, pp. 2584–2591. [Google Scholar]
  15. Basak, A.; Das, S.; Tan, K.C. Multimodal Optimization Using a Biobjective Differential Evolution Algorithm Enhanced with Mean Distance-Based Selection. IEEE Trans. Evol. Comput. 2013, 17, 666–685. [Google Scholar] [CrossRef]
  16. Deb, K.; Saha, A. Multimodal optimization using a bi-objective evolutionary algorithm. Evol. Comput. 2012, 20, 27–62. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, Y.; Li, H.X.; Yen, G.G.; Song, W. MOMMOP: Multiobjective Optimization for Locating Multiple Optimal Solutions of Multimodal Optimization Problems. IEEE Trans. Cybern. 2015, 45, 830–843. [Google Scholar] [CrossRef] [PubMed]
  18. Yao, J.; Kharma, N.; Grogono, P. Bi-Objective Multipopulation Genetic Algorithm for Multimodal Function Optimization. IEEE Trans. Evol. Comput. 2010, 14, 80–102. [Google Scholar]
  19. Yu, W.J.; Ji, J.Y.; Gong, Y.J.; Yang, Q.; Zhang, J. A Tri-Objective Differential Evolution Approach for Multimodal Optimization. Inf. Sci. 2017, 423, 1–23. [Google Scholar] [CrossRef]
  20. Shir, O.M.; Emmerich, M.; Back, T. Self-Adaptive Niching CMA-ES with Mahalanobis Metric. In Proceedings of the IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007. [Google Scholar]
  21. Li, X.; Engelbrecht, A.; Epitropakis, M. Benchmark Functions for CEC 2013 Special Session and Competition on Niching Methods for Multimodal Function Optimization; Technical Report; Royal Melbourne Institute of Technology: Melbourne, VIC, Australia, 2013. [Google Scholar]
  22. Cheng, R.; Li, M.; Li, K.; Yao, X. Evolutionary Multiobjective Optimization Based Multimodal Optimization: Fitness Landscape Approximation and Peak Detection. IEEE Trans. Evol. Comput. 2018, 22, 692–706. [Google Scholar] [CrossRef]
  23. Qu, B.Y.; Suganthan, P.N.; Liang, J.J. Differential Evolution with neighborhood mutation for multimodal optimization. IEEE Trans. Evol. Comput. 2012, 16, 601–614. [Google Scholar] [CrossRef]
  24. SMahfoud, W. Niching Methods for Genetic Algorithms. Ph.D. Thesis, University of Illinois at Urbana-Champaign, Champaign, IL, USA, 1995. [Google Scholar]
  25. Mengsheol, O.; Goldberg, D. Probabilistic crowding: Deterministic crowding with probabilistic replacement. In Proceedings of the Genetic and Evolutionary Computation Conference, Orlando, FL, USA, 13–17 July 1999; pp. 409–416. [Google Scholar]
  26. Fayek, M.B.; Darwish, N.M.; Ali, M.M. Context based clearing procedure: A niching method for genetic algorithms. J. Adv. Res. 2010, 1, 301–307. [Google Scholar] [CrossRef] [Green Version]
  27. Hui, S.; Suganthan, P.N. Ensemble and Arithmetic Recombination-Based Speciation Differential Evolution for Multimodal Optimization. IEEE Trans. Cybern. 2016, 46, 64–74. [Google Scholar] [CrossRef] [PubMed]
  28. Yang, Q.; Chen, W.N.; Yu, Z.; Gu, T. Adaptive Multimodal Continuous Ant Colony Optimization. IEEE Trans. Evol. Comput. 2017, 21, 191–205. [Google Scholar] [CrossRef]
  29. Haghbayan, P.; Nezamabadi-Pour, H.; Kamyab, S. A niche GSA method with nearest neighbor scheme for multimodal optimization. Swarm Evol. Comput. 2017, 35, 78–92. [Google Scholar] [CrossRef]
  30. Qu, B.Y.; Suganthan, P.N.; Das, S. A Distance-Based Locally Informed Particle Swarm Model for Multimodal;Optimization. IEEE Trans. Evol. Comput. 2013, 17, 387–402. [Google Scholar] [CrossRef]
  31. He, X.; Zhou, Y. Enhancing the performance of differential evolution with covariance matrix self-adaptation. Appl. Soft Comput. 2018, 64, 227–243. [Google Scholar] [CrossRef]
  32. Wang, Y.; Li, H.X.; Huang, T.; Li, L. Differential evolution based on covariance matrix learning and bimodal distribution parameter setting. Appl. Soft Comput. J. 2014, 18, 232–247. [Google Scholar] [CrossRef]
  33. Chen, X.; Tianfield, H.; Du, W.; Liu, G. Biogeography-based optimization with covariance matrix based migration. Appl. Soft Comput. 2016, 45, 71–85. [Google Scholar] [CrossRef]
  34. Ghosh, S.; Das, S.; Roy, S.; Sk, M.I. A Differential Covariance Matrix Adaptation Evolutionary Algorithm for real parameter optimization. Inf. Sci. 2012, 182, 199–219. [Google Scholar] [CrossRef]
  35. He, X.Y.; Zhou, Y.R.; Chen, Z.F. Evolutionary Bilevel Optimization based on Covariance Matrix Adaptation. IEEE Trans. Evol. Comput. 2018, 1–21. [Google Scholar] [CrossRef]
  36. Awad, N.; Ali, M.Z.; Suganthan, P.N.; Reynolds, R.G. A novel differential crossover strategy based on covariance matrix learning with Euclidean neighborhood for solving real-world problems. In Proceedings of the IEEE Congress on Evolutionary Computation, San Sebastian, Spain, 5–8 June 2017; pp. 380–386. [Google Scholar]
  37. Jiang, Q.; Wang, L.; Cheng, J.; Zhu, X.; Li, W.; Lin, Y.; Yu, G.; Hei, X.; Zhao, J.; Lu, X. Multi-objective differential evolution with dynamic covariance matrix learning for multi-objective optimization problems with variable linkages. Knowl. Based Syst. 2017, 121, 111–128. [Google Scholar] [CrossRef]
  38. Wang, T.C.; Liaw, R.T.; Ting, C.K. MOEA/D using covariance matrix adaptation evolution strategy for complex multi-objective optimization problems. In Proceedings of the IEEE Congress on Evolutionary Computation, Vancouver, BC, Canada, 24–29 July 2016; pp. 983–990. [Google Scholar]
  39. Wang, J.H.; Liao, J.J.; Zhou, Y.; Cai, Y.Q. Differential Evolution Enhanced with Multiobjective Sorting-Based Mutation Operators. IEEE Trans. Cybern. 2014, 44, 2792–2805. [Google Scholar] [CrossRef] [PubMed]
  40. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  41. Wang, Y.; Cai, Z.X.; Zhang, Q.F. Differential evolution with composite trial vector generation strategies and control parameters. IEEE Trans. Evol. Comput. 2011, 15, 55–66. [Google Scholar] [CrossRef]
  42. Alcalá-Fdez, J.; Sánchez, L.; García, S.; del Jesus, M.J.; Ventura, S.; Garrell, J.M.; Otero, J.; Romero, C.; Bacardit, J.; Rivas, V.M.; et al. KEEL: A software tool to assess evolutionary algorithms to data mining problems. Soft Comput. 2009, 13, 307–318. [Google Scholar] [CrossRef]
Figure 1. Relationship between multimodal optimization in a decision space and multi-objective optimization in an objective space. (a) Three optimal solutions of a multimodal function in a decision space. (b) Three optimal solutions of a transformed multi-objective function in an objective space.
Figure 1. Relationship between multimodal optimization in a decision space and multi-objective optimization in an objective space. (a) Three optimal solutions of a multimodal function in a decision space. (b) Three optimal solutions of a transformed multi-objective function in an objective space.
Algorithms 12 00056 g001
Figure 2. Transformation example of the equal maxima function.
Figure 2. Transformation example of the equal maxima function.
Algorithms 12 00056 g002aAlgorithms 12 00056 g002b
Figure 3. Number of global optima found by the matrix adaptation evolution strategy (MA-ES) over three independent runs on f2.
Figure 3. Number of global optima found by the matrix adaptation evolution strategy (MA-ES) over three independent runs on f2.
Algorithms 12 00056 g003
Figure 4. Equal maxima.
Figure 4. Equal maxima.
Algorithms 12 00056 g004
Figure 5. Himmelblau.
Figure 5. Himmelblau.
Algorithms 12 00056 g005
Figure 6. Shubert 2D function.
Figure 6. Shubert 2D function.
Algorithms 12 00056 g006
Figure 7. Composition Function 2.
Figure 7. Composition Function 2.
Algorithms 12 00056 g007
Figure 8. Box-plot of peaks found by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on 20 test problems.
Figure 8. Box-plot of peaks found by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on 20 test problems.
Algorithms 12 00056 g008aAlgorithms 12 00056 g008b
Figure 9. Average number of peaks found by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO versus the number of FES on eight test problems.
Figure 9. Average number of peaks found by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO versus the number of FES on eight test problems.
Algorithms 12 00056 g009
Table 1. Parameter setting for test functions.
Table 1. Parameter setting for test functions.
Fun.εrDNumber of Global OptimaNumber of Function EvaluationsPopulation Size
f10.1/0.01/0.001/0.0001/0.000010.01125 × 10480
f20.1/0.01/0.001/0.0001/0.000010.01155 × 10480
f30.1/0.01/0.001/0.0001/0.000010.01115 × 10480
f40.1/0.01/0.001/0.0001/0.000010.01245 × 10480
f50.1/0.01/0.001/0.0001/0.000010.5225 × 10480
f60.1/0.01/0.001/0.0001/0.000010.52182 × 105100
f70.1/0.01/0.001/0.0001/0.000010.22362 × 105300
f80.1/0.01/0.001/0.0001/0.000010.53814 × 105300
f90.1/0.01/0.001/0.0001/0.000010.232164 × 105300
f100.1/0.01/0.001/0.0001/0.000010.012122 × 105100
f110.1/0.01/0.001/0.0001/0.000010.01262 × 105200
f120.1/0.01/0.001/0.0001/0.000010.01282 × 105200
f130.1/0.01/0.001/0.0001/0.000010.01262 × 105200
f140.1/0.01/0.001/0.0001/0.000010.01364 × 105200
f150.1/0.01/0.001/0.0001/0.000010.01384 × 105200
f160.1/0.01/0.001/0.0001/0.000010.01564 × 105200
f170.1/0.01/0.001/0.0001/0.000010.01584 × 105200
f180.1/0.01/0.001/0.0001/0.000010.011064 × 105200
f190.1/0.01/0.001/0.0001/0.000010.011084 × 105200
f200.1/0.01/0.001/0.0001/0.000010.012084 × 105200
Table 2. Experimental results of ANP (average peaks found) obtained by the covariance matrix adaptation evolution strategy (CMA-ES), matrix adaptation evolution strategy (MA-ES), CMA-ES-Niching (CMA-N) CMA-ES-Niching with the multi-objective optimization algorithm (CMA-NMO), CMA-ES Niching with Mahalanobis Metric (CMA-NMM), CMA-NMM with multi-objective optimization algorithm (CMA-NMM-MO), MA-ES-Niching (MA-ESN), and matrix adaptation evolution strategy with multi-objective optimization algorithm (MA-ESN-MO) for functions f1f10.
Table 2. Experimental results of ANP (average peaks found) obtained by the covariance matrix adaptation evolution strategy (CMA-ES), matrix adaptation evolution strategy (MA-ES), CMA-ES-Niching (CMA-N) CMA-ES-Niching with the multi-objective optimization algorithm (CMA-NMO), CMA-ES Niching with Mahalanobis Metric (CMA-NMM), CMA-NMM with multi-objective optimization algorithm (CMA-NMM-MO), MA-ES-Niching (MA-ESN), and matrix adaptation evolution strategy with multi-objective optimization algorithm (MA-ESN-MO) for functions f1f10.
Fun.εCMA-ESMA-ESCMA-NCMA-NMOCMA-NMMCMA-NMM-MOMA-ESNMA-ESN-MO
f10.122222222
0.0122222222
0.00122222222
0.0000121.96222222
f20.111555555
0.011155554.925
0.001114.525554.325
0.00001113.203.6851.962.161.44
f30.111111111
0.0111111111
0.0011111110.961
0.00001110.600.2810.480.560.60
f40.111442.68444
0.0111442.68444
0.001112.6442.6842.564
0.00001112.563.962.563.802.363.96
f50.11.12 1.04222222
0.011.121.041.8021.6421.802
0.0011.121.041.5221.4821.602
0.000011.121.041.401.801.321.721.321.80
f60.11113.5617.5612.7217.5214.1217.40
0.011113.4817.4812.7217.4814.0417.36
0.0011113.4417.4012.7217.3614.0017.36
0.0000110.9613.4017.2012.7217.2414.0017.04
f70.11136.0035.2827.6835.0436.0035.20
0.011129.6830.4427.6829.0028.8430.00
0.0011129.4030.2827.6828.9628.8429.64
0.000011129.5229.9227.1628.3228.5629.12
f80.11.080.9642.6453.4044.0454.0443.4450.84
0.011.080.9242.6452.9644.0453.4843.4050.20
0.0011.080.7642.6452.4044.0452.9243.4049.68
0.000011.080.6442.6050.8044.0451.2843.3647.72
f90.111.0421689.2028.9273.3221681.48
0.0111.0031.4057.3228.9252.1630.4854.40
0.00111.0031.4046.9228.9241.6030.4843.72
0.0000110.9631.4033.5628.9231.0030.4831.00
f100.1119.9211.688.6411.729.6811.72
0.01119.9211.608.6411.729.6811.56
0.001119.9211.408.6411.649.6811.28
0.00001119.9211.328.6411.249.6810.72
Table 3. Experimental results of ANP (average peaks found) obtained by CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO for functions f11f20.
Table 3. Experimental results of ANP (average peaks found) obtained by CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO for functions f11f20.
Fun.εCMA-ESMA-ESCMA-NCMA-NMOCMA-NMMCMA-NMM-MOMA-ESNMA-ESN-MO
f110.111.046.004.283.724.166.004.04
0.0111.043.683.963.723.963.724.00
0.00111.043.683.923.723.963.604.00
0.0000111.043.563.883.723.923.363.96
f120.1112.846.922.806.243.007.04
0.01112.766.442.805.602.686.60
0.001112.646.082.805.402.486.08
0.00001112.645.802.805.362.405.80
f130.1115.763.923.443.965.643.92
0.01113.363.843.443.963.443.92
0.001113.323.843.443.963.323.92
0.00001113.163.723.243.923.203.92
f140.110.966.003.561.723.685.803.64
0.0110.761.763.521.723.681.843.60
0.00110.761.763.521.723.681.843.52
0.0000110.721.763.481.723.481.843.48
f150.112.088.002.401.122.447.442.20
0.0110.841.202.241.122.441.322.12
0.00110.801.202.081.122.441.322.12
0.0000110.721.202.001.122.401.322.12
f160.1106.002.281.162.726.002.32
0.01101.242.161.162.441.082.08
0.001101.242.041.162.281.082.00
0.00001101.241.921.162.241.081.84
f170.10.8406.321.7211.525.921.56
0.010.8401.001.5211.441.041.56
0.0010.8401.001.5211.441.041.56
0.000010.8401.001.4811.401.041.48
f180.13.88061.842.321.5261.60
0.010.88011.601.001.4811.60
0.0010.24011.561.001.4811.56
0.000010.16011.161.001.0811.04
f190.10.7201.201.121.081.041.401.08
0.010.2001.001.121.001.001.001.04
0.0010.1601.001.081.000.841.000.96
0.00001001.001.001.000.801.000.80
f200.100817.72180.92
0.01000.880.160.960.360.560
0.001000.0400000
0.0000100000000
Table 4. Experimental results of ANP, Peak ratio (PR) and Success Rate (SR) obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.0001 for functions f1 to f10.
Table 4. Experimental results of ANP, Peak ratio (PR) and Success Rate (SR) obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.0001 for functions f1 to f10.
Fun.CMA-ESMA-ESCMA-NCMA-NMOCMA-NMMCMA-NMM-MOMA-ESNMA-ESN-MO
f1ANP2(≈)1.96(≈)2(≈)2(≈)2(≈)2(≈)2(≈)2
PR10.98111111
SR100%96%100%100%100%100%100%100%
f2ANP1(−)1(−)3.64(−)4.32(≈)5(+)4.28(≈)3.04(−)4.40
PR0.20.20.720.8618.560.680.88
SR0%0%28%40%100%40%4%44%
f3ANP1(≈)1(≈)0.72(−)1(≈)1(≈)1(≈)0.72(−)1
PR110.721110.721
SR100%100%72%100%100%100%72%100%
f4ANP1(−)1(−)2.60(−)3.96(≈)2.56(−)4(≈)2.48(−)4
PR0.250.250.650.990.6410.621
SR0%0%8%96%8%100%4%100%
f5ANP1.12(−)1.04(−)1.52(−)1.92(≈)1.36(−)1.92(≈)1.48(−)1.92
PR0.560.520.760.960.680.960.740.96
SR12%4%60%92%48%92%52%92%
f6ANP1(−)1(−)13.40(−)17.36(≈)12.72(−)17.36(≈)14.00(−)17.20
PR0.050.050.740.960.700.960.770.95
SR0%0%0%52%4%60%0%44%
f7ANP1(−)1(−)29.52(≈)30.00(≈)27.60(−)28.72(≈)28.60(≈)29.20
PR0.020.020.820.830.760.790.790.82
SR0%0%0%0%0%0%0%0%
f8ANP1.08(−)0.68(−)42.60(−)51.60(≈)44.04(−)52.04(≈)43.36(−)48.88
PR0.010.010.520.630.540.640.530.60
SR0%0%0%0%0%0%0%0%
f9ANP1(−)0.96(−)31.40(−)39.68(≈)28.92(−)35.88(≈)30.48(−)36.44
PR0.0040.0040.140.180.130.160.140.16
SR0%0%0%0%0%0%0%0%
f10ANP1(−)1(−)9.92(−)11.40(≈)8.64(−)11.44(≈)9.68(−)11.20
PR0.080.080.820.950.720.950.800.92
SR0%0%0%56%4%60%0%48%
−/≈/+8/2/08/2/08/2/00/10/07/2/10/10/08/2/0\
Table 5. Experimental results of ANP, PR, and Success Rate (SR) obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.0001 for functions f11f20.
Table 5. Experimental results of ANP, PR, and Success Rate (SR) obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.0001 for functions f11f20.
Fun.CMA-ESMA-ESCMA-NCMA-NMOCMA-NMMCMA-NMM-MOMA-ESNMA-ESN-MO
f11ANP1(−)1.04(−)3.56(−)3.92(≈)3.72(−)3.96(≈)3.60(−)4
PR0.160.170.600.650.620.660.600.66
SR0%0%0%0%0%0%0%0%
f12ANP1(−)1(−)2.64(−)6.00(≈)2.80(−)5.40(−)2.40(−)5.96
PR0.120.120.330.750.350.670.310.75
SR0%0%0%16%0%0%0%4%
f13ANP1(−)1(−)3.20(−)3.80(≈)3.44(−)3.96(≈)3.20(−)3.92
PR0.160.160.540.630.570.660.530.65
SR0%0%0%0%0%0%0%0%
f14ANP1(−)0.72(−)1.76(−)3.52(≈)1.72(−)3.64(≈)1.84(−)3.48
PR0.160.120.290.580.280.600.300.58
SR0%0%0%0%0%0%0%0%
f15ANP1(−)0.76(−)1.20(−)2.04(≈)1.12(−)2.40(≈)1.32(−)2.12
PR0.120.090.150.250.140.300.160.26
SR0%0%0%0%0%0%0%0%
f16ANP1(−)0(−)1.24(−)1.96(≈)1.16(−)2.24(≈)1.08(−)1.84
PR0.1600.200.330.190.370.180.32
SR0%0%0%0%0%0%0%0%
f17ANP0.84(−)0(−)1(−)1.48(≈)1(−)1.44(≈)1.04(−)1.48
PR0.1000.120.180.120.180.130.19
SR0%0%0%0%0%0%0%0%
f18ANP0.16(−)0(−)1(≈)1.44(+)1(≈)1.32(+)1(≈)1.08
PR0.0200.160.240.160.220.160.18
SR0%0%0%0%0%0%0%0%
f19ANP0.04(−)0(−)1(≈)1.04(≈)1(≈)0.80(≈)1(≈)0.96
PR0.00500.120.130.120.100.120.12
SR0%0%0%0%0%0%0%0%
f20ANP0(≈)0(≈)0.04(≈)0(≈)0(≈)0(≈)0(≈)0
PR000.00500000
SR0%0%0%0%0%0%0%0%
−/≈/+9/1/09/1/07/3/00/9/17/3/01/8/17/3/0\
Table 6. Results obtained by the Wilcoxon test for algorithm MA-ESN-MO.
Table 6. Results obtained by the Wilcoxon test for algorithm MA-ESN-MO.
VSR+RExact p-ValueAsymptotic p-Value
CMA-ES188.51.5≥0.20.000144
MA-ES208.51.5≥0.20.000103
CMA-N182.08.0≥0.20.00043
CMA-NMO52.5137.5≥0.21
CMA-NMM176.513.5≥0.20.000911
CMA-NMM-MO85.0105.0≥0.21
MA-ESN205.54.5≥0.20.000152
Table 7. Average ranking of the algorithms (Friedman).
Table 7. Average ranking of the algorithms (Friedman).
AlgorithmRanking
CMA-ES6.8
MA-ES7.4
CMA-N4.525
CMA-NMO2.25
CMA-NMM4.75
CMA-NMM-MO2.55
MA-ESN5
MA-ESN-MO2.725
Table 8. Experimental results of the success rate obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.001
Table 8. Experimental results of the success rate obtained by the CMA-ES, MA-ES, CMA-N, CMA-NMO, CMA-NMM, CMA-NMM-MO, MA-ESN, and MA-ESN-MO on ε = 0.001
Fun.CMA-ESMA-ESCMA-NCMA-NMOCMA-NMMCMA-NMM-MOMA-ESNMA-ESN-MO
f1100%100%100%100%100%100%100%100%
f220%20%90.4%100%100%100%86.4%100%
f3100%100%100%100%100%100%96%100%
f425%25%60%100%67%100%65%100%
f562%50%80%100%82%100%84%100%
f65.5%5.3%71.7%98%76%97.1%77.5%97.1%
f72.7%2.7%79%83.3%77.8%80.2%81.5%83%
f81.3%0.7%52.8%64.1%54.6%66.1%51.1%60.9%
f90.4%0.4%14.4%22.2%13.7%17.6%14.6%21.1%
f108.3%8.3%85.3%97%75.6%98%83.6%94.6%

Share and Cite

MDPI and ACS Style

Li, W. Matrix Adaptation Evolution Strategy with Multi-Objective Optimization for Multimodal Optimization. Algorithms 2019, 12, 56. https://doi.org/10.3390/a12030056

AMA Style

Li W. Matrix Adaptation Evolution Strategy with Multi-Objective Optimization for Multimodal Optimization. Algorithms. 2019; 12(3):56. https://doi.org/10.3390/a12030056

Chicago/Turabian Style

Li, Wei. 2019. "Matrix Adaptation Evolution Strategy with Multi-Objective Optimization for Multimodal Optimization" Algorithms 12, no. 3: 56. https://doi.org/10.3390/a12030056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop