Introduction

Many-objective optimization problems (MaOPs) refer to those multi-objective optimization problems with more than three objectives [1, 2]. Interestingly, most MaOPs in the real world are problems with irregular Pareto fronts [3, 4]. The Pareto optimal solutions of these problems are distributed only in part of the objective space. Several characteristics make such problems more difficult to solve than problems with regular Pareto fronts [5, 6]. Firstly, since the distribution of Pareto front is unknown, it is hard to identify the real Pareto front area. The dominated individuals affect the judgment and waste computing resources, thus slowing down convergence, and even misleading the convergence direction. Secondly, since the Pareto optimal solutions are only distributed in part of the objective space, it is difficult to randomly generate enough individuals in the Pareto area. Thirdly, the irregular shape of the Pareto front makes it more difficult to partition the objective space uniformly, especially in high-dimensional space, which brings difficulties to decomposition-based algorithms.

In recent years, MaOPs with irregular Pareto fronts have received increasing attention in the field of evolutionary optimization. Several evolutionary algorithms have been developed for this kind of problems. These algorithms can be roughly divided into two categories. Firstly, algorithms based on clustering or space division. In our recent work [6], a clustering based multi-objective evolutionary algorithm called CA-MOEA was proposed, which adaptively generates a set of cluster centers in the current population as the reference points to maintain diversity and accelerate convergence for problems with irregular Pareto fronts. However, as CA-MOEA is mainly based on non-dominated sorting, it encounters resistance in dealing with many-objective problems. MaOEA/C [7] partitions the population by two clustering approaches and selects individuals with best convergence to form the final population. However, due to frequent changes in clustering results, population convergence will be affected. RdEA [8] predicts the geometry of the Pareto front according to the distribution of the population, and proposes a region division method to partition the objective space. In each region, individuals closer to the reference point or the center point of the region are selected. [9, 10]are also algorithms based on division. Secondly, algorithms based on adjustment of reference vectors. RVEA* [11], a variant of RVEA [11], was proposed to deal with irregular Pareto fronts. In this algorithm, if no individual is associated with a reference vector, it will be replaced by a randomly generated vector. Consequently, the convergence directions varies with iteration and it also difficult to guarantee an even partition of the population, resulting in wasted computing resources and deteriorated convergence performance. The very recently proposed iRVEA [12] suggested a new vector adaption method to further improved the performance of RVEA on irregular problems. A-NSGA-III [13] is a variant of NSGA-III [13] for irregular Pareto fronts, reference points with which no individual is associated will be deleted and relocated around a reference point associated with more than two individuals. The algorithm can easily be misled by individuals in incorrect directions at the beginning of iteration. Meanwhile, such repeated detection, deletion and relocation operations are inefficient and leads to a high computational complexity. Furthermore, the original reference points need to be kept for diversity considerations, requiring additional computation resources. Similar to other algorithms, relocating reference points may cause useless calculation and an uneven distribution of the solutions. There are several variants of MOEA/D [14] for problems with complex Pareto fronts. For example, MOEA/D-AWA [15] produces new weight vectors by non-dominated solutions found by now to adjust the weight vectors periodically for finding the appropriate convergence direction. Self-organizing map (SOM) is applied in MOEA/D-SOM [16] to adjusting the weight vectors for MOEA/D. The problem is that training a SOM network costs a lot of calculation and the performance of SOM depends on the distribution of initial random individuals which probably not reflecting the true distribution of the Pareto front. Besides, MaOEA/D-2ADV [17], MOEA/D-SAS [18], and MOEA/D-AM2M [19] all make efforts to get reasonable results on problems with irregular Pareto fronts. CLIA [20] use incremental Support Vector Machine (SVM) to judge the effectiveness of reference vectors and adjust the distribution of reference vectors. Obviously, training classifier requires a lot of computation. Moreover, as dominated individuals usually do not reflect the distribution of Pareto fronts, the training results have a risk of error. Meanwhile, changing the reference vectors time to time will affect the speed of convergence.

Especially, if an M-objective optimization problem has a Pareto front of a dimension less than (M-1), the Pareto front is called a degenerate Pareto front [21] which is an important type of irregular Pareto front. In particular, DTLZ5 and DTLZ6 with more than three objectives are a class of degenerate problems that are especially hard for MOEAs to converge to the Pareto front. To the best of our knowledge, no existing algorithms are able to obtain well converged solutions on DTLZ5 and DTLZ6 problems with more than three objectives.

This work aims to develop an effective evolutionary algorithm tailored for solving a class of challenging MaOPs with degenerated Pareto fronts, such as DTLZ5 and DTLZ6 with more than three objectives. To this end, a many-objective evolutionary algorithm based on Pareto front location and piecewise mapping clustering (called LC-MaOEA) is proposed. In LC-MaOEA, the potential Pareto front regions are detected and a set of reference vectors based on clustering of projected solutions are produced within these regions to maintain stable and reliable convergence. To be specific, LC-MaOEA can be seen a decomposition-based evolutionary algorithm, in which the population is guided by four sets of reference vectors, namely, location vectors, axis vectors, mapping-based cluster center vectors and in some cases, Gaussian random vectors. These vectors cooperate together to maintain diversity and speed up convergence. Meanwhile, neighborhood-based mating selection mechanism is used to improve the ability of the algorithm to produce more individuals in Pareto front area.

The main contributions of this work can be summarized as follows:

  1. 1)

    A novel evolutionary algorithm LC-MaOEA is proposed. In this algorithm, a new vector-based Pareto front location method is designed, which improves the utilization of computing resources and makes the population converge to the Pareto front area more efficiently. The proposed algorithm can solve the DTLZ5 and DTLZ6 problems with more than three objectives very well, and has achieved good results in dealing with similar practical application problems.

  2. 2)

    A new method for generating reference vectors is proposed. Four kinds of vectors are generated by this method. Among them, location vectors locate the approximate position of Pareto front, axis vectors are used to find the boundary individuals, and cluster center vectors are generated by clustering the individuals mapped to the piecewise hyperplanes to improve diversity. Finally, a set of Gaussian random vectors will be generated within the Pareto front area if the number of cluster center vectors is not enough.

  3. 3)

    The proposed LC-MaOEA is examined on DTLZ5 and DTLZ6 problems with 5 to 40 objectives and compared with four state-of-the-art evolutionary algorithms designed for irregular Pareto fronts. A widely used performance indicators are adopted to evaluate the performance of LC-MaOEA. Furthermore, LC-MaOEA is applied to optimize the polyester filament melt-transportation process. Our experimental results demonstrate that LC-MaOEA is promising for solving this kind of real world problems.

Basic ideas

Pareto front location

The Pareto front of DTLZ5 and DTLZ6 with more than three objectives is distributed in a narrow area of the objective space. Figure 1 shows the Pareto front of 3-objective DTLZ5 [22] problem which is a typical problem with degenerate Pareto front. The hollow circles are reference points produced by Das and Dennis’s systematic approach [23], and the reference vectors are generated by the origin and these reference points. The solid points are Pareto optimal solutions of the 3-objective DTLZ5 problem, the stars are effective vectors with the smallest angle to at least one Pareto optimal solution. Figure 2 shows the dominated solutions of the 3-objective DTLZ5 problem. The hollow circles and the stars have the same meaning as in Fig. 1, but the solid points represent the dominated individuals in one of the early iteration. It can be seen from Fig. 1 that in a degenerate problem, a very small proportion of reference vectors are effective. The remaining vectors not only occupy computing resources, but also mislead the convergence direction. That is why algorithms using fixed reference vectors are difficult to perform well in such problems.

Fig. 1
figure 1

Pareto front of 3-objective DTLZ5 problem

To remedy this, some algorithms adjust the reference vectors according to the distribution of current population. However, as shown in Fig. 2, since the distribution of dominated population do not accurately reflect the real distribution of Pareto front, those ineffective reference vectors will lead to an ineffective area, thus population cannot converge to the correct PF. Moreover, since population changes from generation to generation, the frequently changing reference vectors affects convergence rate.

Fig. 2
figure 2

Dominated solutions of 3-objective DTLZ5 problem

Based on the above considerations, in this paper, we proposed a vector-based detection method to locate the Pareto front and providing a reliable convergence direction for the population. Firstly, we set a number of uniformly distributed reference vectors by Das and Dennis’s systematic approach [23]. Note that these reference vectors are not only set for detection, but also working as the convergence direction in the early iteration. Meanwhile, every individual will be assigned to a vector which has the minimum angle with it, and those vectors that have no less than one individual assigned to them is remarked as possible effective vectors. After that, we evaluate the change of possible effective vectors, and if the distribution of possible effective vectors tends to be stable, we get the final set of effective vectors which will be treated as location vectors. The location vectors will keep guiding the population until the end of the iteration. This method clarifies the convergence direction and speeds up the convergence rate.

Piecewise mapping-based cluster center vectors

Decomposition-based evolutionary algorithm has been regarded as a promising approach for handling MaOPs. These algorithms decompose a problem into a number of sub-problems by a set of vectors. In this paper, we decided to apply this approach. So we need a set of direction vectors that can uniformly partition the degenerate Pareto front. Although we get some location vectors by vector-based effective direction detection method, the number of location vectors in the degenerate Pareto front problem is far from enough to maintain the diversity of the population, so the auxiliary reference vectors is necessary. As mentioned in our previous work CA-MOEA [6], hierarchical clustering is useful to produce reference points in irregular shaped area. However, in many-objective optimization problems, as there are so many non-dominated individuals in the objective space, we need more powerful convergence method. As shown in Figs. 3 and 4a–j are ten individuals in the objective space, LV1 and LV2 are two location vectors. If we do hierarchical clustering directly among the population, individuals with the least sum of squared errors to each other will be merged into one cluster [6], so A, B and C are merged into one cluster, D is a separate cluster, E and G are merged into a new cluster, F and H are merged into one cluster, I and J are merged into a cluster, just as shown in Fig. 3. Thus we get five cluster centers by calculating the average objective values of members in each cluster. Then we can calculate five cluster center vectors CV1, CV2, CV3, CV4 and CV5 by original point and five cluster centers. It can be seen from Fig. 3, on one hand, as cluster A, B and C, cluster F and H have cluster centers too close to each other, CV2 and CV3 almost coincided. On the other hand, although I and J are very close to each other, they are actually in two completely different convergence direction. Moreover, although D is far away from E, G and I, actually they belong to the same convergence direction. Obviously, as this clustering method cannot distinguish the convergence direction of individuals, it cannot appropriately partition the convergence directions. Besides, we have also tried some other clustering methods, including angle-based clustering and so on, to find suitable partitions, but the results are not good enough.

Fig. 3
figure 3

Cluster center vectors produced by clustering directly

For better partitioning, we propose a new mapping-based clustering method. Firstly, the hyper planes are calculated with the location vectors as the normal vectors. In Fig. 4, the two solid line segments represent the hyper planes with location vectors LV1 and LV2 as the normal vectors, separately. Secondly, we do mapping for each individual which are associated with location vectors. To do this, we use the origin and each individual to make a straight line, and then calculate the intersection of the straight line and the hyperplane. This intersection is the mapping point on the hyperplane of an individual. Thirdly, we do hierarchical clustering on the hyper planes using the Ward’s linkage criterion measured by cosine distance. Fourthly, cluster centers are calculated in every cluster, and cluster center vectors are calculate by these cluster centers and the origin. By this way, individuals in the near convergence direction will be clustered into a cluster, just like the cluster consist of A and F, and the cluster constructed by D, E, G and H. Finally, we get more uniformly distributed reference vectors CV1, CV2, CV3, CV4 and CV5. These cluster center vectors cooperate with location vectors to guiding the convergence direction of population while maintaining diversity.

Fig. 4
figure 4

Cluster center vectors produced by clustering the mapped candidates

Proposed algorithm LC-MaOEA

In this section, we present the details of LC-MaOEA. Similar to MOEA/D [14], each individual of LC-MaOEA corresponds to a reference vector. Then, the neighborhood of each reference vector is calculated, and offspring is generated by the individuals with the neighborhood serial numbers. The PBI value is used to evaluate the individuals and do the environmental selection. A detection method is used in LC-MaOEA to predict the effective convergence directions of the Pareto optimal individuals. Thus, the environmental selection in LC-MaOEA differs from that in MOEA/D which uses fixed reference vectors generated by Das and Dennis’s systematic approach. In the following, we first present the framework of LC-MaOEA, followed by a description of the main components of the algorithm.

LC-MaOEA

Figure 5 presents the algorithm flowchart of LC-MaOEA. At first, an initial parent population of size N is randomly generated according to the problem to be solved. Then, N uniformly distributed vectors V are generated by Das and Dennis’s systematic approach [23]. Meanwhile, h is initialized as 0 that indicates the Pareto area has not been detected. After that, the closest b vectors of each vector are calculated as its neighbor vectors and the ordinal numbers of the neighbor vectors are stored in B. To improve the utilization of the vectors, the objective values of each individual are scaled as follows:

$$\begin{aligned} P_{i,n}^j = P_i^j - z_{\min }^j, \end{aligned}$$
(1)
Fig. 5
figure 5

Cluster center vectors produced by clustering the mapped candidates

where, \(P_i^j\) represents the jth objective of individual \({P_i}\), \(z_{\min }^j\) represents the minimum value in the jth objective of all the individuals. Then, to locate the Pareto front, every scaled individual is assigned to a vector with the smallest angle to itself, and the vector has at least one individual assigned to it is regarded as effective vector in the current iteration. When the effective vectors remain unchanged in several iterations, it is said that the effective vectors are stable. At this time, h will be set as 1 to avoid duplicate detection.

The stable effective vectors will be regarded as location vectors LV, which is used to locate the approximate position of Pareto front. Then the axis vectors \([0,\ldots ,0,1],[0,\ldots ,0,1,0],\ldots ,[1,0,\ldots ,0]\) are used to search boundary individuals. Since the Pareto front is degenerate, the number of LV is so limited that the population diversity cannot be guaranteed. So the cluster center vectors CV are designed to fix this issue.

The way to produce cluster center vectors CV is explained in Algorithm1. In addition, if the number of direction vectors is less than population size N, we generate some random vectors within the range of effective directions to complement the number of reference vectors. For generating offspring, each individual \({P_i}\) will randomly choose one of its neighbor individuals from B(i), and generate a new individual \({Q_i}\) by simulated binary crossover (SBX) [24] and polynomial mutation [25] with a crossover probability \({p_\mathrm{{c}}} = 1.0\) and a mutation probability \({p_\mathrm{{m}}} = 1/D\), where, D is the number of decision variables. Here, we use the sequence number of the vector, not the sequence number of the individual itself. Since at first the vectors and individuals do not correspond one to one, this mating pattern is equivalent to the random mating pattern, which is conducive to exploring more areas of the objective space. Then, as the iteration progresses, each vector will find the individual with smaller PBI [14] values, which make it becomes a neighborhood-based enhanced mating model. In our opinion, this method is more suitable for handling many-objective problems with degenerate Pareto fronts. Because it not only has certain exploratory ability, but also can concentrate to produce more individuals in effective directions. In environmental selection, the new produced individual will be compared with each neighbor parent and the one with smaller PBI value will be selected. This main loop repeats until a termination condition is met and the population produced by the last generation will be the final population.

Cluster center vector generation

The cluster center vectors are designed to better partition the convergence direction so as to maintain the diversity of the population. As explained in Algorithm 1, for each location vector \({{\text {LV}}_i}\), the hyperplane \({H_i}\) with \({{\text {LV}}_i}\) as normal vector is calculated. Thus we obtain several piecewise hyperplanes. After that, all the individuals \(P_{{\text {LV}}_i}\) assigned to \({{\text {LV}}_i}\) are mapped to the hyperplane \({H_i}\), and be recorded as \(P_{{\text {map}}_i}\). All the non-duplicate mapped individuals \(P_{{\text {map}}_i}\) are merged to K clusters by hierarchical cluster method. Ward’s linkage criterion measured by cosine distance is used to do the hierarchical clustering. The parameter K is no more than the number of non-duplicate mapped individuals \(P_{{\text {map}}_i}\). Afterwards, the clustering center of each cluster is obtained by the average value of each objective of all individuals in the cluster. Detailed hierarchical clustering method and clustering center calculation method can refer to CA-MOEA [6]. Finally, the cluster vectors CV are calculated by the origin and these cluster center points. Since this clustering method is able to evenly divide the mapped individuals into any number of groups, the clustering center vectors are able to evenly partition the convergence direction.

figure a

Computational complexity of LC-MaOEA

The worst-case complexity of the major operations in one generation of the proposed LC-MaOEA is analyzed as follows. First, the objective normalization requires O(MN). Next, assigning N individuals to N vectors takes O(N). For calculating cluster center vectors, mapping takes O(N), calculating cluster centers takes \(O({N^2})\). Then, generating offspring requires O(N). For environmental selection, the comparison of PBI value takes O(bN), b is the size of neighbors which is smaller than N.

To summarize, the computational complexity of LC-MaOEA in the worst cases is \(O({N^2})\), which is comparable to MOEA/D-AM2M, RVEA*, A-NSGA-III and CA-MOEA.

Empirical results

In this section, the performance of LC-MaOEA is empirically examined on widely used scalable benchmark problems DTLZ5, DTLZ6 which have degenerate Pareto fronts when the objective number is more than or equal to three. Meanwhile, LC-MaOEA is compared with four state-of-the-art algorithms designed for irregular Pareto front, namely, MOEA/D-AM2M, RVEA*, A-NSGA-III and CA-MOEA. All experiments are run on PlatEMO [26].

Performance indicators

The inverted generational distance (IGD) [27] is adopted as performance indicator in this work, which is calculated as follows:

$$\begin{aligned} {\text {IGD}}(\mathbf{{A}},{\mathbf{{Z}}_{\text {eff}}}) = \frac{1}{{\left| {{\mathbf{{Z}}_{eff}}} \right| }}\sum \limits _{i = 1}^{\left| {{\mathbf{{Z}}_{\text {eff}}}} \right| } {{\min }_{j = 1}^{\left| \mathbf{{A}} \right| } } \mathrm{{ }}d({\mathbf{{z}}_i},{\mathbf{{a}}_j}), \end{aligned}$$
(2)

where \(\mathbf{{A}} = [{\mathbf{{a}}_1},{\mathbf{{a}}_2},\ldots ]\) is the non-dominated individuals obtained by an MOEA in the objective space, \({\mathbf{{Z}}_{\text {eff}}} = [{\mathbf{{z}}_1},{\mathbf{{z}}_2},\ldots ]\) is a set of solutions sampled from the known Pareto front. \(d({\mathbf{{z}}_i},{\mathbf{{a}}_j}) = {\left\| {{\mathbf{{z}}_i} - {\mathbf{{a}}_j}} \right\| _2}\) is the Euclidean distance from \({\mathbf{{z}}_i}\) to \({\mathbf{{a}}_j}\). The IGD metric is able to measure both diversity and convergence of A if \(\left| {{\mathbf{{Z}}_{\text {eff}}}} \right| \) is large enough. \(\left| {{\mathbf{{Z}}_{\text {eff}}}} \right| \) here is set to 10,000 for 2 and 3 objective problems. Note that a smaller IGD value indicates a better performance. As the degenerate front distributes in a very narrow area of the objective space, and the widely used HV indicator [28] is calculated by hypervolume, so there will be a mistake that poorly converged populations have better HV values. Based on above consideration, we don’t use HV indicator to measure the results.

Table 1 the statistical results (mean and standard deviation) of IGD values on 5–40 objectives DTLZ5 and DTLZ6 problems calculated by MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA
Fig. 6
figure 6

a Pareto optimal solutions of 10-objective DTLZ6 and the optimization results of 10-objective DTLZ6 obtained by b MOEA/D-AM2M, c RVEA*, d A-NSGA-III, e CA-MOEA, f LC-MaOEA

Experimental settings

The algorithms are implemented in MATLAB R2014a and run on a PC with Intel Core i7-6500U, CPU 2.50 GHz, RAM of 12.00 GB. The population size is set according to \(N = C_{H + m - 1}^{m - 1}\) [18], where, H specifies the granularity or resolution of weight vectors by Das and Dennis’s systematic approach, m represents the number of objectives. Considering the number of objectives and the parameter settings in the existing literature, we set the population size to 126 for 5-objective problems, 275 for 10-objective problems, 240 for 15-objective problems, 230 for 20-objective problems, 465 for 30-objective problems and 820 for 40-objective problems. We do not test the problems with more than 40 objectives, but we think the algorithm can solve these problems with more than 40 objectives. The maximum iteration numbers are set to 250 for DTLZ5 and DTLZ6. These parameter settings are the same as recommended in [29] and [30]. The simulated binary crossover (SBX) [24] and polynomial mutation [25] are adopted in LC-MaOEA with a crossover probability \(p_\mathrm{{c}} = 1.0\) and a mutation probability \(p_\mathrm{{m}} = {1 / D}\), where, D represents the number of decision variables. Meanwhile, the distribution index of crossover is set to and the distribution index of mutation, as recommended in [30].

IGD results on the test suites

Table 1 present the IGD results of MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA on test problems with 5–40 objectives. The IGD values are averaged over 20 independent runs. The mean values for each instance are presented and the standard deviation are listed. The results obtained by LC-MaOEA are shaded if they are the best among the five compared algorithms. For more rigorous comparisons, the Wilcoxon rank sum test is used to compare the result of the algorithms at a significance level of 0.05. Symbol ‘+’ represents that LC-MaOEA significantly outperforms the compared algorithm, while symbol ‘−’ indicates the compared algorithm performs significantly better than LC-MaOEA. Meanwhile, symbol ‘\(\approx \)’ states there is no significant difference between these two algorithms.

It can be found from Table 1 that LC-MaOEA achieves the best performance on all the twelve test problems of DTLZ5 and DTLZ6. It is noteworthy that our algorithm improves the results of these two problems by more than one order of magnitude. Moreover, MOEA/D-AM2M reaches the second best results on eleven out of twelve test problems of DTLZ5 and DTLZ6, which shows its ability on handling this kind of problems.

In addition, the performance of RVEA* and A-NSGA-III on these test problems are mediocre. On the one hand, RVEA* achieves the second best result on 5-objective DTLZ6 problem and the third place on rest five test problems of DTLZ6. It also reaches the third place on all the DTLZ5 test problems. On the other hand, A-NSGA-III get the fourth place on all the test problems. These results may due to the fact that these two algorithms are constantly adjusting the convergence direction according to the current population, so that they are easy to be deceived, and their exploration ability and convergence pressure are insufficient.

Figure 6 shows the results of 10-objective DTLZ6 problems by Pareto optimal solutions, MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA. Abscissa of the figures is the objective order number and longitudinal coordinates are the objective values of each individual on each objective. Each large graph has the same coordinate range. It should be noted, however, that since the range of the longitudinal coordinates of Pareto front and LC-MaOEA are relatively small, we have enlarged the effective area with a small graph with the range of the longitudinal coordinates from 0 to 1. It can be seen from Fig. 6 that our proposed LC-MaOEA can well approximate the Pareto front of 10-objective DTLZ6 problem, while none of the other algorithms can converge near the Pareto front. But among them, we can see that MOEA/D-AM2M find the most Pareto optimal individuals, although there are still many dominated individuals. Apart from LC-MaOEA, the results of RVEA* converge in the range closest to Pareto front, nevertheless, it can find very few Pareto optimal individuals so that the Pareto front area is almost blank. Moreover, the results of A-NSGA-III are even worse which can be attributed to the ineffective update mechanism of the weight vectors and the slow convergence rate due to frequent changes in the reference direction.

The fast non-dominated sorting based CA-MOEA performs the worst among the five compared algorithms in that the final population are far away from the PF. The reason why these comparison algorithms cannot simulate Pareto front of DTLZ6 well is that DTLZ6 is a highly misleading nonlinear conflict problems [22], so these algorithms lack strong recognition and convergence ability to achieve satisfactory results. Overall, it can be seen that LC-MaOEA shows significant better capability in approximating the Pareto fronts of DTLZ5 and DTLZ6 problems compared to the other state-of-the-art algorithms.

Comparison on convergence

We further compare the performance of MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA by examing their convergence speed.

Fig. 7
figure 7

Convergence rate on a 10-objective DTLZ5, b 15-objective DTLZ6

Figure 7 demonstrates the curves of IGD value over iterations on (a) 10-objective DTLZ5, (c) 15-objective DTLZ6, respectively. The red circles denote the change in IGD values of LC-MaOEA, and the black points, crosses, stars, triangles denote that of MOEA/D-AM2M, RVEA*, A-NSGA-III, and CA-MOEA, respectively. As shown in Fig. 7, LC-MaOEA converges faster in the earlier generations on all of these test problems with degenerate Pareto fronts, and then stayed at a stable level, which owe to its stable convergence direction.

Optimization of the polyester filament melt-transportation process

Polyester filament has the characteristics of high strength, good heat resistance, good elasticity, wrinkle resistance, wear resistance and good insulation. It has been widely used in various fields of national economy. Melt-transportation is an important process in polyester filament production. The schematic diagram of melt transportation process is demonstrated in Fig. 8.

Fig. 8
figure 8

Schematic diagram of polyester filament melt-transportation process

According to [31], five performance indexes are selected as the optimization objectives, namely, the intrinsic viscosity \({f_1}\), temperature \({f_2}\), pressure of the melt in spinneret inlet pipe \({f_3}\), the residence time of the melt during the whole process \({f_4}\) and the pressure of the melt in metering pump inlet pipe \({f_5}\). Six decision variables affect the objective function, namely, initial intrinsic viscosity of the melt \({\text {IV}}_0\), initial temperature \({T_0}\), initial pressure \({P_0}\), initial flow rate G, temperature of the heat-transfer medium pump \({T_{\text {sb}}}\), and the small heat-transfer pump used in the melt-transportation process \({T_{\text {ss}}}\). The objective function formulas are given in (3).

$$\begin{aligned} \left\{ \begin{array}{l} \min \mathrm{{ }}{f_1}(\mathbf{{x}}) = \left| {{y_{\varDelta IV}} - \varDelta IV} \right| \\ \min \mathrm{{ }}{f_2}(\mathbf{{x}}) = \left| {{y_{\varDelta T}} - \varDelta T} \right| \\ \min \mathrm{{ }}{f_3}(\mathbf{{x}}) = \left| {{y_{\varDelta P}} - \varDelta P} \right| \\ \min \mathrm{{ }}{f_4}(\mathbf{{x}}) = \left| {{y_{{P_{\text {b}}}}} - {P_{\text {b}}}} \right| \\ \min \mathrm{{ }}{f_5}(\mathbf{{x}}) = {y_{\text {time}}} \end{array} \right. \end{aligned}$$
(3)

where, \(\mathbf{{x}} = (I{V_0},{T_0},{P_0},G,{T_{\text {sb}}},{T_{\text {ss}}})\), \(y_{{\varDelta \text {IV}}}\) represents mechanism model output for reducing melt intrinsic viscosity, \(y_{\varDelta T}\) is the increase of the melt temperature during the whole process, \({y_{\varDelta P}}\) represents mechanism model output for reducing melt pressure, \(y_{{P_\text {b}}}\) represents the mechanism model output of the pressure of the melt in spinneret inlet pipe, \(y_{{\text {time}}}\) represents the output of the melt residence time of the whole process. Meanwhile, \(\varDelta IV\), \(\varDelta T\), \(\varDelta P\), \({P_\text {b}}\) represent the expectation value of \(y_{\varDelta \text {IV}}\), \({y_{\varDelta T}}\), \(y_{\varDelta P}\), \(y_{{P_\text {b}}}\), respectively. This is a 5-objective optimization problem. We compare the performance of MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA. The parameter set are the same as suggested in [31]. Figure 9 shows the results of melt-transportation process obtained by reference points given by [31], MOEA/D-AM2M, RVEA*, A-NSGA-III, CA-MOEA and LC-MaOEA. Since the range of vertical coordinates of each result is too different, we have not unified them. So as shown in Fig. 9, the maximum objective value of MOEA/D-AM2M is 5, RVEA* is 7, A-NSGA-III is 8, CA-MOEA is 25, the reference points and LC-MaOEA is 0.8. We can see that LC-MaOEA can best simulate the reference results. What’s more, according to [31], the smaller the results, the better. As the results of LC-MaOEA concentrate in the smallest range, it quite meet the requirement.

Fig. 9
figure 9

a Reference points of melt-transportation process and the optimization results obtained by b MOEA/D-AM2M, c RVEA*, d A-NSGA-III, e CA-MOEA, f LC-MaOEA

In conclusion, LC-MaOEA performs well on real-world application with irregular Pareto front which are difficult to converge.

Conclusion

In this paper, a many-objective evolutionary algorithm based on Pareto front location and piecewise mapping clustering (LC-MaOEA) is proposed for solving a class of MaOPs with degenerate Pareto fronts. The main idea of the proposed algorithm is to guide the population to converge along more efficient and promising directions. To achieve this, a Pareto front detection method is proposed to find the reliable convergence direction, and a mapping-based hierarchical clustering algorithm is developed to better partition the convergence direction. Meanwhile, a neighbor-based mating selection method is used to produce more individuals in the effective convergence directions. The proposed LC-MaOEA is compared with four state-of-the-art algorithms designed for degenerate problems, and tested on DTLZ5 and DTLZ6 problems with 5–40 objectives. In addition, LC-MaOEA is applied to optimize the parameters in polyester filament melt-transportation process. The results on both benchmark problems as well as the application demonstrate that LC-MaOEA is able to obtain satisfactory solutions on MaOPs with degenerate Pareto fronts.

Despite its attractive performance demonstrated on both test problems and a real-world application, LC-MaOEA has much room for improvement in that it is competitive mainly for a particular class of MaOPs with degenerated problems. Our future work includes extending LC-MaOEA to solve problems with various shapes of Pareto fronts.