PBI function based evolutionary algorithm with precise penalty parameter for unconstrained many-objective optimization

https://doi.org/10.1016/j.swevo.2019.100568Get rights and content

Abstract

Fixed or experiential penalty parameter of the penalty-based boundary intersection (PBI) function method cannot simultaneously ensure the convergence and diversity for all shape of Pareto front (PF). Too large penalty parameter may lead to bad convergence while too small parameter can not ensure the diversity. Specially, if the penalty parameter is too small, some reference weight vectors may have no solution on it. This error is hard to be rectified. In this paper, we prove that the lower bound of the penalty parameter is determined by three factors. The first one is the shape of the PF. The second one is the cosine distance between two adjacent reference vectors. The third one is the number of objectives. We deduce the lower bound of the penalty parameter. Once the penalty parameter was calculated, an individual with minimal PBI function is attached to the corresponding reference vector. The minimal-PBI-function-first principle is used in the environmental selection to guarantee the wideness and uniformity of the solution set. The time complexity is low. The proposed method is compared with other three state-of-the-art many-objective evolutionary algorithms on the unconstrained test problems MaOP, DTLZ and WFG with up to fifteen objectives. The experimental results show the competitiveness and effectiveness of the proposed algorithm in both time efficiency and accuracy.

Introduction

Multi-Objective Problems (MOPs) with more than three objectives are informally known as Many Objective Optimization Problems (MaOPs) [1,2]. Generally, a MaOP with only box constraints can be stated as follows.minimizeF(x)=(f1(x),,fm(x))Tsubject ​to ​xΩ

ΩRn is the decision (variable) space, x=(x1,,xn)TΩ is a candidate solution. Here n is the number of decision variables, and m is the number of objectives. F:ΩRm constitutes m objective functions, and Rm is the objective space. The attainable objective set is defined as Θ={F(x)|xΩ}. The x(1) is said to dominate x(2) (denoted as x(1)x(2)) iff for i{1,,m}fi(x(1))fi(x(2)) and j{1,,m} fj(x(1))<fj(x(2)). A solution x* is a Pareto-optimal solution to formula (1) if there is no other solution xΩ such that xx. F(x*) is then referred to as Pareto-optimal (objective) vector. The set of all Pareto-optimal solutions is referred to as the Pareto-optimal set (PS). Accordingly, the set of all Pareto-optimal vectors,  ​PF={F(x)Rm|xPS} is referred to as the Pareto front (PF) [11].

Recently, MaOPs are extensively researched within the Evolutionary Multi-objective Optimization (EMO) community. A number of Many-Objective Evolutionary Algorithms (MOEAs) have been proposed to deal with MaOPs [[4], [5], [6], [7], [8], [9], [10], [11]]. Based on the Pareto optimization and dominance, lots of MOEAs are proposed. The improved Strength Pareto Evolutionary Algorithm (SPEA2) [12] is proposed by Zitzler et al. The Nondominated Sorting Genetic Algorithm II (NSGA-II) is proposed in Ref. [13]. Besides, the Pareto Archived Evolution Strategy (PAES) [14] and Pareto envelope-based selection algorithm II (PESAII) [15] belong to such category. Due to the obstacles caused by the Dominance Resistance (DR) and the Active Diversity Promotion (ADP) phenomena [4], the Pareto based methods encounter difficulties when handling MaOPs. The DR phenomenon refers to the incomparability of solutions in terms of the Pareto dominance relation. The main reason is that the proportion of non-dominated solutions in a population rise rapidly as the number of objectives increases [16]. The probability of any two solutions being comparable is η = 1/2m−1 in an m-dimensional objective space [17]. When m = 10, η is as low as 0.002. As a result of DR phenomenon, the dominance-based primary selection criterion fails to distinguish between solutions, and then the diversity based secondary selection criterion is activated to determine the survival of solutions in the environmental selection. This is the so-called ADP problem which may be harmful to the convergence of the approximated Pareto fronts. Furthermore, for the 10-objective problem, the diversity maintenance mechanism even makes the population gradually move away from the true Pareto front. To address the DR phenomenon, a number of modified or relaxed Pareto dominance relations were proposed to enhance the selection pressure toward the true Pareto front. The typical methods include the ε-dominance [18], α-dominance [19], fuzzy Pareto dominance [20], L-dominance [21], θ-dominance [6], dominance area control [22], grid-dominance [23] and preference order ranking [24].

Another category of MOEA is the decomposition-based method, which decomposes a MOP into a set of subproblems and optimizes them in a collaborative way. The MOEA/D proposed in Ref. [25] is a representative of this category. In this algorithm, a predefined set of weight vectors are used to specify search directions towards different parts of the PF. In MOEA/D, each solution is associated with a subproblem, and each subproblem is optimized by using information from its neighborhoods. Since the weight vectors of subproblems are widely distributed, the obtained solutions are expected to have a wide spread over the PF. The original study on MOEA/D investigated three decomposition methods namely—the WS, the weighted TCH and the penalty-based boundary intersection (PBI).

The penalty-based boundary intersection(PBI) approach used in literatures [[25], [26], [27], [28], [29]] is defined as follows.Minimizeg(x|w,z)=d1+θd2Subject ​to ​xΩwhered1=(F(x)z)Twwd2=F(x)(z+d1ww)where z=(z1,z2zm)T is ideal objective vector with zi=minxΩfi(Ω). Fig. 1 presents an example to illustrate d1 and d2 of a solution x with regard to a weight vector w=(0.5, 0.5)T. The d1 is used to evaluate the convergence of x toward the PF, and d2 is a measure of population diversity. g(x|w,z)=d1+θd2 plays as a composite measure of both convergence and diversity. The goal of PBI approach is to push F(x) as low as possible so that it can reach the boundary of the PF [29]. The major shortcomings of this approach are their applicability to MaOPs [26] and that it may result in low population diversity in some problems [27,28].

Sato [26] extends the conventional PBI approach and proposes inverted PBI (IPBI) decomposition method. In the conventional decomposition methods, such as the TCH and the PBI, solutions are evolved toward the reference point by minimizing the scalarizing function value. On the contrary, in the IPBI approach, solutions are evolved from the nadir point by maximizing the scalarizing function value [28]. The experimental study on MOKPs and WFG4 problem [30] with 2–8 objectives illustrates that the IPBI approach can better approximate widely spread PF in comparison to other scalarizing approaches. Just like the PBI, the limitation of the IPBI approach is that it involves the parameter θ which needs to be appropriately tuned. Yang et al. [31] suggest two penalty schemes, referred to as adaptive penalty scheme (APS) and subproblem-based penalty scheme (SPS), to set the value of the parameter θ. In the APS, θ is linearly increased with the number of generations from θmin (=1) to θmax (=10). A small value of θ is adopted initially so as to emphasize convergence and drive the search toward PF as fast as possible. The value of θ is gradually increased so as to emphasize diversity toward the later search stage. One of the main drawbacks of the PBI decomposition approach is that there is no unique setting of the penalty parameter θ that works well on different types of problems with different number of objectives [32].

As pointed out by Anupam Trivedi et al. [28], introduction of new methods to adaptively control parameter θ can be interesting research direction. This is what we do in this paper. The basic idea of the proposed PBI function based method comes from (1) Now that d1 evaluates the convergence of x toward the PF, the time consuming Pareto domination calculation is not necessary; (2) Now that d2 governs the diversity, the niche preservation operator and neighborhood concept for mating restriction are not necessary. Only if the value of θ is properly set, the convergence and diversity can be ensured. We provide a mathematic foundation of θ. Without the dominance calculation, the proposed method is more efficient and accurate. The major contributions of this paper are as follows.

  • 1)

    Precise penalty parameter calculation

The value of θ depends on the shape of PF and the number of objectives. We provide the mathematic foundation of θ calculation. This method applicable to various types of problems with different number of objectives. The calculation is simple and easy to use.

  • 2)

    Stable normalization

Traditionally, the nadir point (z1max,,zmmax) is derived within the combination (of parent and child population) set. The zimax denotes the extreme value of the ith objective. In the child population, there might be ‘bad’ individual which is far away from PF and has extraordinary large value of objective. It makes the search process unstable [32]. We know that the ‘bad’ individuals are eliminated after the environmental selection. Hence, the nadir and points are calculated after the environmental selection. These points are used in the next iteration of the MOEA. Besides, a damping mechanism is introduced into the normalization calculation. Initially, the normalization is less sensitive to the ideal and nadir points so as to emphasize convergence and drive the search toward PF as fast as possible. With the increase of evolutionary generations, the normalization is more and more sensitive to the ideal and nadir points so as to emphasize diversity.

This paper is organized as follows. The mathematic foundation of θ calculation is illustrated in Section 2. Our proposed algorithm is descripted in Section 3. Experimental settings and comprehensive experiments are discussed in Section 4. Finally, Section 5 concludes this paper.

Section snippets

PBI function contour

Two-objective example is used to illustrate the contour calculation. In Fig. 2, the angle between the weight vector w and the horizontal axis is α, 0oα90o, β=90oα.

A known objective point F(x0) is on the weight vector w. Its PBI function value g(x0) is calculated by formula (2). There are two contours starting from point F(x0). One intersects horizontal axis on point F(x1) = [f1(x1), 0]. Another one intersects vertical axis on point F(x2) = [0, f2(x2)]. Because that F (x0) and F(x1) are all

Proposed algorithm: Moea/4P

The proposed PBI function with precise penalty parameter (MOEA/4P) method is shown in Algorithm 1. The initialization procedure generates N initial solutions and H reference weight vectors. The reference weight vectors generation method comes from Das and Dennis’s literature [35]. As discussed in Ref. [2], in order to have intermediate weight vectors within the simplex, we should set p ≥ m, where p is the amount of divisions along each objective. However, in a high-dimensional objective space,

Experimental study

We compare the proposed algorithm with MOEA/DD1 [33], NSGA-III [2] and VaEA2 [1] on test suits DTLZ [38], WFG [39] and MaOP [40]. The proposed MOEA/4P and NSGA-III are realized in MATLAB. Java code of NSGA-III can also be downloaded from internet.3

Conclusion and future work

Too large θ has negative effect on convergence while too small θ results in bad diversity. Some researches notice that the performance of the decomposition-based methods rely on the shape of PF. Some other researches furtherly find that the performance of PBI based methods (which belongs to the decomposition-based category) strongly depend on the θ value. To the best of our knowledge, the precise θ calculation method was not researched. Especially, the mathematic foundation of the θ value is

Acknowledgment

The authors would like to thank the anonymous reviewers for their insightful comments which help to improve the quality and readability of this paper.

References (41)

  • M. Li et al.

    Bi-goal evolution for many-objective optimization problems

    Artif. Intell.

    (2015)
  • Hui Li et al.

    Comparison between MOEA/D and NSGA-III on a set of novel many and multi-objective benchmark problems with challenging difficulties

    Swarm Evol. Comput.

    (2019)
  • Y. Xiang et al.

    A vector angle based evolutionary algorithm for unconstrained many-objective optimization

    IEEE Trans. Evol. Comput.

    (2016)
  • K. Deb et al.

    An evolutionary many-objective optimization algorithm using reference-point-based nondominated sorting approach, Part I: solving problems with box constraints

    IEEE Trans. Evol. Comput.

    (2014)
  • S. Yang et al.

    A grid-based evolutionary algorithm for many-objective optimization

    IEEE Trans. Evol. Comput.

    (2013)
  • M. Li et al.

    Shift-based density estimation for pareto-based algorithms in many-objective optimization

    IEEE Trans. Evol. Comput.

    (2014)
  • Z. He et al.

    Visualization and performance metric in many-objective optimization

    IEEE Trans. Evol. Comput.

    (2016)
  • Y. Yuan et al.

    A new dominance relation-based evolutionary algorithm for many-objective optimization

    IEEE Trans. Evol. Comput.

    (2016)
  • Y. Yuan et al.

    Balancing convergence and diversity in decomposition-based many-objective optimizers

    IEEE Trans. Evol. Comput.

    (2016)
  • C. Zhu et al.

    Generalization of pareto-optimality for many-objective evolutionary optimization

    IEEE Trans. Evol. Comput.

    (2016)
  • M. Elarbi et al.

    A new decomposition-based NSGA-II for many-objective optimization

    IEEE Trans. Syst., Man, Cybern.: Syst.

    (2017)
  • T. Santos et al.

    On the performance degradation of dominance-based evolutionary algorithms in many-objective optimization

    IEEE Trans. Evol. Comput.

    (2017)
  • E. Zitzler et al.

    SPEA2: Improving the Strength Pareto Evolutionary Algorithm

    (2001)
  • K. Deb et al.

    A fast and elitist multiobjective genetic algorithm: NSGA-II

    IEEE Trans. Evol. Comput.

    (2002)
  • J. Knowles et al.

    The Pareto archived evolution strategy: a new baseline algorithm for Pareto multiobjective optimisation

  • D.W. Corne et al.

    PESA-II: region-based selection in evolutionary multiobjective optimization

  • B. LI et al.

    Many-objective evolutionary algorithms: a survey

    ACM Comput. Surv.

    (2015)
  • M. Laumanns et al.

    Combining convergence and diversity in evolutionary multiobjective optimization

    Evol. Comput.

    (2002)
  • K. Ikeda et al.

    Failure of pareto-based MOEAs: does non-dominated really mean near to optimal?

  • Z. He et al.

    Fuzzy-based pareto optimality for many-objective evolutionary algorithms

    IEEE Trans. Evol. Comput.

    (2014)
  • Cited by (9)

    • Vertical distance-based clonal selection mechanism for the multiobjective immune algorithm

      2021, Swarm and Evolutionary Computation
      Citation Excerpt :

      Recently, a number of multiobjective evolutionary algorithms (MOEAs) have been proposed with promising performance in tackling different kinds of MOPs [4–6]. Generally, according to the selection mechanisms, most existing MOEAs can be classified into three main categories: Pareto-based MOEAs [7–10], decomposition-based MOEAs [11–15], and indicator-based MOEAs [16–20]. In particular, some MOEAs [21–24], which mainly aim to improve the ability of the algorithm to maintain diversity, have shown promising performance.

    • A many-objective algorithm based on staged coordination selection

      2021, Swarm and Evolutionary Computation
      Citation Excerpt :

      In the framework of decomposition, the aggregation function is usually used to replace the Pareto dominance relation. However, different aggregation functions can affect the performance of the algorithms, such as the PBI approach [5] can fluctuate the performance of the algorithm by different penalty parameter θ [27]; the Weighted Sum approach [5] can get excellent performance when the optimization problems have linear PFs, but the performance greatly drops when the optimization problems have concave or convex PFs. In addition, since the decomposition-based method divides the objective space through the pre-defined uniformly distributed reference vectors, the final optimization result of the algorithm also depends on the shape of the true PF.

    View all citing articles on Scopus

    This work was supported in part by the Open Foundation of Guangxi Key Laboratory of Automatic Detecting Technology and Instruments (Grant No. YQ18207). This work was also supported by National Natural Science Foundation of China under Grant No. U1830133.

    View full text