DDR PDF
DDR PDF
DDR PDF
producing a uniformly distributed population when the PF significantly due to the lose of population diversity.
is not linear [21]. To solve the low-efficiency problem of Recent studies have proposed two approaches to alleviate
HV calculation, computationally efficient indicators such as this problem. The first approach is to employ multiple sets
R2 [22] and ∆p [23] have been studied and used in indicator- of reference vectors or multiple aggregation functions. For
based MaOEAs [24], [25]. However, a pre-defined set of example, [40]–[42] suggest to use two aggregation functions
reference points or a subset of the real PF is usually required, simultaneously, one for pulling the solutions toward the ideal
which is also a challenging task. Recently, state-of-the-art point, and the other one for pushing the solutions away
MaOEAs such as BiGE [26], SRA [27], and 1by1EA [28] from the nadir point. Similarly, [43] utilizes two sets of
generally make use of multiple simple indicators to measure reference vectors, one for achieving fast convergence, and
the convergence and diversity separately. Showing significant the other one for approximating a more complete PF. One
performance improvements, these algorithms introduce some main disadvantage of this approach is that it usually requires
problem-dependent parameters which should be tuned care- a specially designed mechanism to control the interactions
fully. between these multiple sets of reference vectors or multiple
As another non-Pareto-based algorithm, the multiob- aggregation functions. The second approach is the reference
jective evolutionary algorithm based on decomposition vector adaptation. One simple implementation is to employ
(MOEA/D) [29] decomposes an MOP into a number of a fixed number of reference vectors and relocate them on
single-objective optimization problems (called subproblems). the fly [35], [44]. To achieve a fast migration to desired
Usually, a pre-defined set of well-spread reference vectors is distribution, [43], [45]–[48] further suggest to adaptively insert
employed to facilitate this decomposition. A neighborhood “useful” reference vectors and delete “non-useful” ones. These
structure is also constructed with the help of these reference adaptation techniques require problem-dependent parameters
vectors. Then, a subproblem is optimized by using the infor- and extra computational burden to calculate the “usefulness”
mation mainly from its neighbor subproblems. for the reference vectors. Besides, a pool containing the
Though designed mainly for solving MOPs, the decom- candidate reference vectors is usually incorporated to facilitate
position strategy used in MOEA/D has been considered as the insertion and deletion. However, how to maintain this
a promising tool in solving MaOPs. One main approach in pool itself is exactly an optimization problem which should
recent studies focuses on improving the diversity manage- be carefully handled.
ment mechanism in existing MOEAs. These algorithms (e.g., Aiming to deal with the above problems, a dynamical
NSGA-III [30], I-DBEA [31], and MOEA/D-DU [32]) try to decomposition strategy and a ranking method is proposed in
measure the population diversity using the distance between this paper. Being different from existing approaches, decom-
the solutions and the reference vectors. Since the reference position in this study does not rely on pre-defined reference
vectors are well-spread, it helps to maintain the population vectors. Detailed properties of this approach are summarized
diversity in high-dimensional objective space. Decomposition as follows.
strategy can also be used to partition the objective space into • No extra pre-defined set of reference vectors are needed.
small subregions. Good solutions are emphasized by using Contrarily, the solution themselves are considered as
traditional methods in each subregion to maintain a balance reference vectors.
between convergence and diversity [33]. Representative ex- • Apart from some common parameters (e.g., the popula-
amples of this type include MOEA/DD [34], RVEA [35], θ- tion size and termination condition), no other parameters
DEA [36], and SPEA/R [37]. are required.
The set of reference vectors plays a key role in • Being robust in coping with problems with complicated
decomposition-based MaOEAs. However, how to set these PFs.
reference vectors is still an open question. In the early studies • An MaOEA is constructed based on the proposed ranking
of MOEA/D, the reference vectors are produced using the method. Also, this ranking method can serve as a sec-
systematic approach [38]. The number of obtained reference ondary criterion in traditional Pareto-based algorithms.
vectors is a combinatorial number determined by the number In the remainder of this paper, we first provide the ba-
of objectives and the number of divisions along each objective. sic ideas of the dynamical decomposition strategy and the
It means this number will increase significantly as the objective dynamical-decomposition-based ranking method (DDR for
number increases. To fix this problem, a two-layer method [30] short) in Section II. The detailed implementations of DDR
is adopted to produce reference vectors in the boundary and and two examples of using DDR in MaOEAs are provided in
inside layer, separately. This method makes the number of Section III. Then, we present the simulation results on a set
reference vectors acceptable in most scenarios. In [37], it is of test instances in Section IV. Finally, Section V concludes
further extended to generate reference vectors with arbitrary this paper and gives some remarks for future studies.
number of layers. Nevertheless, the number of layers is still
an external parameter required to be tuned. II. BASIC I DEAS
Another big problem concerning the reference vectors is that In this section, we first outline the basic idea of the
the performance of decomposition-based MaOEAs strongly dynamical decomposition strategy. Three concepts are then
depends on the shape of PF [39]. They perform well only presented because they are essential in further explanation of
when the distribution of the reference vectors is consistent our ranking method. Finally, we describe the whole process
with the PF shape. Otherwise, their performance deteriorates and the properties of DDR.
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 3
S3. Find a promising solution. Find an unranked solution s ever, since all solutions have been normalized in this study, the ideal point
from A which optimizes the aggregation function: is exactly the origin point. Hence, it is not employed in calculating the
aggregation function.
2 This example assumes that all solutions have positive objective values. This
s = arg min g(x|λ(p)), (6) can be achieved with the normalization procedure described in Section II-B.
x∈A
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 4
Note that, Eq. (5) can be considered as a centroid- Using a set of pre-defined reference vectors in
based clustering process though it is not performed explicitly decomposition-based algorithms has a disadvantage that
throughout the algorithm. This process yields |Q| + 1 clusters some reference vectors may have no intersection with the
of solutions, where the cluster centers are the solutions in PF. Optimizing with these reference vectors will contribute
Q ∪ {p}. A is exactly one of the resulting clusters whose nothing but waste the computational burden. This issue is
center is p. It is the reason why p is utilized to construct the well-settled in the above method since λ(x) always intersects
reference vector in Eq. (6). In the above example (Fig. 1), with the approximated PF at F 0 (x) for all x. Given that
x1 , x2 , and x3 are the three cluster centers. Each of the solutions are getting closer and closer to the PF in the
other solutions is assigned to one of the clusters by checking evolving population, their corresponding reference vectors are
its minimum distance to the three cluster centers, and we very likely to have intersections with the true PF.
obtain three clusters including {x3 , x4 , x7 , x8 }, {x2 , x6 }, and Also, all these reference vectors are located on the hyper-
{x1 , x5 , x9 }. Here, the first cluster is the set A while the last plane f10 + f20 + · · · + fm
0
= 1. This design is based on the
two form the complementary set P − A. Also, pursing the assumption that the problem PF in the normalized objective
diversity performance is the only one criterion in forming the space is this hyperplane. And we hope all solutions will be
candidate set A. In this way, the proposed approach adopts pulled towards this hyperplane along the directions of these
the “diversity first and convergence second” strategy [51]. reference vectors. Note that, this assumption is reasonable if
In traditional decomposition-based algorithms, the objective we lack a priori knowledge of the PF shape.
space is decomposed into many subregions simultaneously.
This kind of decomposition only relies on the reference C. Distance Measure
vectors. Since the reference vectors are pre-defined in most Since we assume the problem PF is a hyperplane in the
algorithms, the produced subregions are always fixed. On the normalized objective space, the distance between two solu-
contrary, in our strategy, the objective space is decomposed tions is measured by the Euclidean distance between their
successively and a new subregion is produced in each decom- corresponding reference vectors:
position. Moreover, the new subregion relies on the subregions
produced previously and are affected by the population solu- dst(x, y) =k λ(x) − λ(y) k2 , (9)
tions. Since the decomposition results in this approach are where k · k2 denotes the 2-norm of a vector.
not determined by the fixed reference vectors, we call it the According to Eq. (9), the distance between two solutions
dynamical decomposition. is only determined by their directions rather than their √ real
positions. Besides, its value is within the range of [0, 2]
B. Reference Vectors regardless of the dimensionality of the space. These properties
The dynamical decomposition strategy relies on reference make it scalable in high-dimensional objective space. In this
vectors when comparing multiple solutions (as shown in step respect, it is very similar to the angular distance. Angular
S3 in Section II-A). Considering the nature of evolutionary distance is widely used in MaOEAs since it provides nice per-
algorithms, the population can be treated as an approximation formance in high dimensional space compared with Euclidean
to the PF. This leads to an idea that we can consider solution distance [15], [28], [35], [37]. In fact, if the reference vectors
themselves as reference vectors. Details are stated as follows: in our proposed approach are located in the hypersphere
We firstly normalize each solution x in P so that their (f10 )2 + (f20 )2 + · · · + (fm
0 2
) = 1, the above distance could be
objective values are inside the range [0, 1]: transformed to the angular distance. This also demonstrates a
fact that using the angular distance gives an assumption that
fi (x) − zimin the normalized PF should be a hypersphere. If this assumption
fi0 (x) = , i = 1, 2, . . . , m (7)
zimax − zimin is not satisfied, the population diversity may not be achieved.
where zimin = min fi (x) and zimax = max fi (x). To prevent
x∈P x∈P
the denominator in Eq. (7) from becoming zero in the case
of zimax = zimin , we fix fi0 (x) = 1e − 10 if zimax − zimin <
1e − 10. Note that, this method is chosen due to its simplicity.
Other methods can also be used to normalize the objective
values. For more details of the normalization mechanisms and
their effects on the algorithm performance, interested readers
may refer to [52].
Then, for each solution x, we calculate its corresponding
reference vector (denoted by λ(x)) using the following for-
mulation:
1 (a) (b)
λ(x) = m 0 F 0 (x), (8)
Σj=1 fj (x) Fig. 2. Well-spread solutions (circles) measured by (a) the angular distance
and (b) the proposed distance.
0
where F (x) = (f10 (x), f20 (x), . . . , fm
0
(x))T
is the normal-
ized objective vector. That is, λ(x) points from the origin Fig. 2 provides an illustration of the difference between
point 0 to F 0 (x). the proposed distance and the angular distance when the PF
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 5
is convex. The thin black curl depicts the normalized PF. The After the initialization, we select a solution s from W using
empty circles denote some solutions located on the PF and the the method proposed in Section II-A. Once s is selected in
red arrows describe the corresponding vectors. In Fig. 2a, each the r-th iteration, it gets the rank r and is moved to Q. These
pair of adjacent solutions has the same angular distance (e.g., operations are executed repeatedly until W is empty.
hx1 , x2 i = hx2 , x3 i, where h·, ·i denotes the acute angle). To have a clear understanding of the process of DDR, Fig. 3
However, it is observed that these solutions are not uniformly provides an illustration of its first three iterations. The circles
distributed over the PF. For example, the distance between depict the solutions in the whole population P . Among them,
x1 and x2 is much larger than that between x2 and x3 . the blue ones depict the ranked solutions, the green ones depict
In Fig. 2b, the blue line depicts the hyperplane f10 + f20 = 1 the pivot p, and the red ones depict the solution s. If a circle
and the blue points depict the reference vectors. Each pair is half green and half red, it means this solution is selected
of adjacent reference vectors has the same Euclidean distance as p and s at the same time. Light yellow areas describe the
(e.g., kλ(x1 ) − λ(x2 )k2 = kλ(x2 ) − λ(x3 )k2 ). It is found subregions (i.e., SA ) and the arrowed dashed lines describe the
that the distribution of x1 , x2 , and x3 is more uniform than reference vectors (i.e., λ(p)). At the beginning, x1 and x2 get
that shown in Fig. 2a. Therefore, the proposed distance rather the rank 0 since they are extreme solutions. Then, at the first
than the angular distance is utilized in this study. iteration (shown in Fig. 3a), x3 is considered as the pivot since
it has the largest distance to x1 and x2 . The subregion SA is
D. Extreme Solutions determined with these solutions. It is found that there are 3
solutions located in this subregion and x4 is the one optimizing
The proposed approach starts with identifying the extreme the corresponding aggregation function. Hence, x4 is chosen
solutions in the population for each axis. This part is not the and gets rank 1. In the next iteration (shown in Fig. 3b), x5
main contribution of this paper, and thus the method proposed is used as the pivot and helps to find the optimum solution
in [30] is employed here. For the j-th axis, its extreme solution x6 which gets the rank 2. In the third iteration, x7 is selected
ej is the one minimizing the aggregation function g(x|w) with to be the pivot, and it is found that itself is the optimum in
w = (w1 , · · · , wm )T being the axis direction: the subregion. Thus, x7 is chosen and ranked. After these
three iterations, totally five solutions x1 , x2 , x4 , x6 , and x7
ej = arg min g(x|w), (10) are ranked with the corresponding ranking results being 0, 0,
( x∈P 1, 2, and 3, respectively.
1 i=j
wi = , i ∈ {1, 2, · · · , m} (11)
1e − 6 i 6= j F. Explanations of DDR
where g is the same aggregation function as that in Eq. (6). Two properties of the ranking results obtained by DDR are
explained as follows:
1) Extreme solutions always obtain the best ranking results:
E. Ranking the Solutions
The extreme solutions play a key role in preserving the
Based on the dynamical decomposition strategy, the ranking population diversity. First, with an appropriate aggregation
method DDR is proposed in this subsection. Suppose P is a function, the extreme solutions are usually the minimum in
population containing N solutions. The task of DDR is to each objective, which determines the ideal point used in the
assign a rank value to each solution in P . The rank value of normalization. Second, they are on the boundary of the PF and
a solution is inside the range {0, 1, . . . , N − 1} and describes have a disproportionate contribution to capturing the whole
its quality in terms of both convergence and diversity. In PF compared to internal solutions [47], [53]. Third, they
this study, a smaller rank value is desirable and the best form the initial set of Q and influence the decomposition
solutions obtain the rank 0. Thus, when applying DDR to of the objective space. Hence, all extreme solutions obtain
evolutionary algorithms, solutions with smaller rank values the rank 0. Note that, in finding the extreme solutions, the
are more likely to survive in the mating selection and the calculation of Eq. (10) are independent of the other solutions.
environmental selection. This makes the overall performance of DDR less influenced by
DDR works like a selection operator which selects unranked the initialization though the reference vectors constructed in
solutions iteratively. It maintains, in its main loop, two sets Q the early stage of the search process may be badly distributed.
and W which consist of the ranked solutions and unranked so- Another concern about the extreme solutions in DDR is that
lutions, respectively. At the beginning, Q and W are initialized they are obtained through searching along the axis directions.
so that Q is the set of all extreme solutions e1 , e2 , . . . , em , It implicitly assumes that the normalized PF is a hyperplane
while W contains all remaining solutions: which is equally inclined to all objective axes and intersects
Q = {e1 , e2 , . . . , em }, W = P − Q. (12) each axis at 1. In other words, DDR treats the normalized PF
as a simplex. One may wonder whether DDR still works if
Then all extreme solutions in Q get the rank 0.3 the problem PF is of non-simplex shape.
Fig. 4 provides an example for the three-objective DTLZ1−1
3 On some problems especially those with degenerate PFs, searching along
problem (discussed in Section S-I-A in the supplementary
different axis directions with Eq. (10) may produce duplicated extreme
solutions. Nevertheless, the algorithm provided in this study still works since material), whose PF is an inverted simplex. The solid black
it does not rely on the number of different extreme solutions in Q. lines depict the assumed simplex PF, which is a triangle with
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 6
(i)
s(i) unless s(j) ∈/ SA . Otherwise, s(j) rather than s(i) will
optimize Eq. (6) in the i-th iteration. In another word, s(j) is
less helpful in improving the convergence performance without
decreasing diversity. Since preserving the diversity is the first
criterion of DDR to rank solutions, it is reasonable that s(i)
obtains a smaller rank value compared with s(j) .
This property can also be illustrated from another aspect.
As mentioned before, we can consider DDR as a selec-
tion operator: iteratively select the best solution from the
unranked ones. In the above example, since Q(i) ⊂ Q(j) ,
distance(x, Q(j) ) ≤ distance(x, Q(i) ) holds for any x.
(i) (j)
Consequently, SA is very likely to be larger than SA .
When solutions are well-distributed in the objective space,
A(i) will contain more candidate solutions than A(j) does.
Fig. 4. Extreme solutions on the 3-objective DTLZ1−1 problem.
It means the selection pressure is usually large in the previous
stage, but is gradually decreased as more and more solutions
are selected and ranked. From this aspect, previously ranked
its apexes at (1, 0, 0)T , (0, 1, 0)T and (0, 0, 1)T . The green solutions are of high quality because they have survived in
lines depict the true PF after normalization, which is a rotated fierce competitions in the selection. Thus, they obtain smaller
triangle. We also assume that all solutions located on the true rank values compared with those ranked in the later stage.
PF have already been added into the current population P
and the task of DDR is to find out a finite set of uniformly III. D ETAILED I MPLEMENTATIONS
distributed solutions. Obviously, the true extreme solutions In this section, the detailed implementations of DDR are
are g1 , g2 , and g3 . But in DDR, e1 , e2 , and e3 are regarded presented in the form of pseudo-code. We also provide two
as the extreme ones and receive the rank 0. Through some simple examples to show how to use DDR in MaOEAs.
elementary geometric computations, it is found that g1 , g2 ,
and g3 are the midpoints of the boundaries of the assumed PF A. Implementations of DDR
while e1 , e2 , and e3 are the midpoints of the boundaries of the
true PF. Also, g1 , g2 , and g3 are the solutions furthest away All the details of DDR have been discussed in Section II.
from {e1 , e2 , e3 }. Hence, g1 , g2 , and g3 will be chosen as the However, a simple trick can be used to reduce the computa-
promising solutions during the first three iterations of DDR. tions in Eqs. (1) and (5) based on the following fact:
Their rank values can only be chosen from {1, 2, 3} which are distance(x, Q ∪ {s}) = min{distance(x, Q), dst(x, s)}.
also relatively small (good) rank values. This example shows (13)
that DDR still works in finding the extreme solutions even This means we can calculate distance(x, Q) by iteratively
when the problem PF is of non-simplex shape. updating. For convenience, we introduce a new symbol md(x)
2) Previously ranked solutions obtain smaller rank values (means the minimum distance to the ranked solutions) to
than those ranked subsequently: With slightly abusing sym- denote distance(x, Q). Once a solution s is ranked and its
(r)
bols, we denote s(r) , Q(r) , A(r) , and SA as the symbols corre- distance to x is smaller than md(x), md(x) is updated.
sponding to s, Q, A, and SA in the r-th iteration, respectively. Algorithm 1 provides the pseudo-code of DDR. π(x) de-
Given i < j and i, j ∈ {1, . . . , N }, s(j) cannot dominate notes the ranking result of the solution x. Lines 1-4 perform
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 7
is performed to partition the population into multiple fronts and DTLZ4 especially in high-dimensional objective space.
(i.e., F1 , F2 , . . . ). Each non-dominated front is selected one at RVEA is outperformed by both DDEA and DDEA+NS on
a time to construct the new population unless the size of new more than 23 test instances, but it performs pretty well on
population exceeds N . Also, solutions in each front are ranked WFG5, WFG9, and DTLZ3. NSGA-III, MOMBI-II, and VaEA
using DDR. If the solutions in the i-th front can not be added also achieve competitive results on high-dimensional DTLZ
into the population but the population size is smaller than N , problems. GDE3 and MOEA/D do not perform well and are
DDR is performed in the i-th front. Then, the best N − |P | surpassed by both DDEA and DDEA+NS on almost all the
solutions according to the DDR ranking results are selected test instances.
from the i-th front and appended to P . It is worth noting that Here, we discuss the correlations between the characteristics
since DDR employs the aggregation function to increase the of the problems and the algorithm performance. GDE3 and
convergence, the proposed algorithms DDEA and DDEA+NS MOEA/D are not especially designed for solving MaOPs, so
are also considered as aggregation-based MaOEAs in essence. their performance are relatively poor compared with other
competitors when m ≥ 5. For example, GDE3 fails to
C. Time Complexity converge to the PF region on high-dimensional DTLZ1 and
DTLZ3 as it produces a zero HV value. It is because its dif-
In DDR, identifying the extreme solutions and initializ- ferential evolution (DE) operator is likely to generate offspring
ing the md values (line 6 and line 12 in Algorithm 1) distant from their parents, which is seen as undesirable in
takes O(N m2 ) time. Dynamical decomposition (lines 16-18 the context of many-objective optimization [30], [37], [63].
in Algorithm 1) requires O(N m) computations for ranking a On the other hand, the performance of the other competi-
solution. After ranking a solution, updating the md values for tors seem to be problem-dependent. For example, they have
the unranked solutions also takes O(N m) time. When m < N , similar performance on relatively easier problems such as
the overall time complexity for ranking all the population DTLZ1, DTLZ2, and DTLZ4. However, on DTLZ3 which
solutions in DDR is O(N 2 m). has a huge number of local optima, algorithms using pre-
DDEA only relies on DDR to rank the solutions. Thus, defined reference vectors perform significantly better than the
the time complexity of one generation of DDEA is O(N 2 m). others. The underlying reason is that DTLZ3 has a regular PF
Whereas, in DDEA+NS, extra O(N log m−2 N ) computations whose shape is consistent with the distribution of the reference
are required in the non-dominated sort [54]. The rankings vectors. Pulling the solutions to the ideal point along the
executed in all fronts require O(N 2 m) in the worst case. directions determined by these reference vectors will preserve
Hence, the worst-case time complexity of one generation in both convergence and diversity. Without using the pre-defined
DDEA+NS is max{O(N 2 m), O(N log m−2 N )}. reference vectors, DDEA, DDEA+NS, and VaEA are faced
with difficulties in maintaining the diversity and are likely
IV. C OMPARATIVE S TUDIES to be trapped into local optima. This phenomenon is also
To demonstrate the effectiveness of the dynamical de- observed on WFG5 and WFG9 since they are highly deceptive
composition strategy, we compare DDEA and DDEA+NS and a set of pre-defined reference vectors is crucial to preserv-
with six state-of-the-art algorithms including GDE3 [55], ing the population diversity. On the other problems, DDEA
MOEA/D [29], NSGA-III [30], MOMBI-II [35], RVEA [35], and DDEA+NS possess substantial advantages. Particularly,
and VaEA [15]. 30 widely used benchmark problems are though their performance on DTLZ2 is not outstanding, DDEA
selected, namely WFG1 to WFG9 [56], DTLZ1 to DTLZ4 and DDEA+NS are superior to the other competitors on MaF2.
and DTLZ7 [57], DTLZ1−1 to DTLZ4−1 [39], MaF2, MaF5, As MaF2 is intended to increase the convergence difficulty of
MaF8, and MaF9 [58], and CPFT1 to CPFT8 [59]. Apart from DTLZ2, it may be concluded that DDEA and DDEA+NS show
the above benchmark problems, Crashworthiness Design of advantages in terms of convergence.
Vehicles is chosen to test the algorithm performance in engi- To visually understand the experimental results, Fig. 5 plots
neering application. Hypervolume (HV) [60] is employed to the final solutions for all algorithms with respect to the 15-
measure both convergence and diversity of the obtained solu- objective WFG5. Both MOEA/D and MOMBI-II perform
tions. Additionally, the averaged Hausdorff distance (∆p ) [23], poorly on this problem and fail to cover more than ten
generational distance (GD) [61] and spread (∆) [6] are chosen objectives. Other algorithms show high performance in terms
as the assistant indicators. All experiments are made using the of coverage. GDE3 reaches the PF region and obtains a
open source software PlatEMO [62]. The detailed experimen- perfectly-distributed solution set. But its solutions may be not
tal settings are provided in the supplementary material. actually close to the PF, which explains why it has a small HV
value. The solutions obtained by DDEA and DDEA+NS are
uniformly distributed on all objectives. Solutions of NSGA-
A. Comparisons on Problems with Regular Pareto Fronts III and RVEA are only located near the PF boundaries. This
Table I summaries the HV results of all algorithms on the phenomenon is resulted from that in the high-dimensional
problems with regular PFs. Generally, DDEA, DDEA+NS, objective space, the reference vectors generated using the
and RVEA present a clear advantage over the other com- systematic approach [38] or its two-layer version [30] are
petitors on the majority of the test instances. DDEA and very likely to be distributed near the boundaries. VaEA covers
DDEA+NS show good performance mainly on WFG4 to all the objectives and its solutions are well-distributed in the
WFG8 and MaF2. They also perform well on DTLZ1, DTLZ2, middle area of the PF. An interesting phenomenon is observed
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 9
TABLE I
M EDIAN AND IQR OF HV RESULTS ON PROBLEMS WITH REGULAR PF S . T HE BEST AND THE SECOND BEST RESULTS FOR EACH TEST INSTANCE
ARE SHOWN WITH DARK AND LIGHT GRAY BACKGROUND , RESPECTIVELY.
TABLE II
M EDIAN AND IQR OF HV RESULTS ON PROBLEMS WITH IRREGULAR PF S . T HE BEST AND THE SECOND BEST RESULTS FOR EACH TEST INSTANCE ARE
SHOWN WITH DARK AND LIGHT GRAY BACKGROUND , RESPECTIVELY.
(a) DDEA (b) DDEA+NS (c) GDE3 (a) DDEA (b) DDEA+NS (c) GDE3
(d) MOEA/D (e) NSGA-III (f) MOMBI-II (d) MOEA/D (e) NSGA-III (f) MOMBI-II
DDEA+NS are proposed. It is demonstrated by the experimen- [17] L. M. S. Russo and A. P. Francisco, “Quick Hypervolume,” IEEE
tal results that DDEA and DDEA+NS are quite competitive Transactions on Evolutionary Computation, vol. 18, no. 4, pp. 481–502,
Aug. 2014.
with the state-of-the-art algorithms on the majority of the test [18] J. Bader and E. Zitzler, “HypE: An Algorithm for Fast Hypervolume-
instances. Moreover, since no reference vectors are employed, Based Many-Objective Optimization,” Evolutionary Computation,
the performance of DDEA and DDEA+NS is less dependent vol. 19, no. 1, pp. 45–76, Mar. 2011.
[19] N. Beume, B. Naujoks, and M. Emmerich, “SMS-EMOA: Multiobjec-
on the PF shapes and is robust especially in solving problems tive selection based on dominated hypervolume,” European Journal of
having irregular PFs. Operational Research, vol. 181, no. 3, pp. 1653–1669, Sep. 2007.
In the future, we will have a further investigation into [20] K. Bringmann and T. Friedrich, “Approximating the volume of unions
and intersections of high-dimensional geometric objects,” Computational
the reference vector generation mechanism in the dynamical Geometry, vol. 43, no. 6-7, pp. 601–610, Aug. 2010.
decomposition strategy. Identifying the dominance resistant [21] M. Emmerich, N. Beume, and B. Naujoks, “An EMO Algorithm Using
solutions may be a promising work to further improve the the Hypervolume Measure as Selection Criterion,” in Evolutionary
Multi-Criterion Optimization. Berlin, Heidelberg: Springer Berlin
algorithm performance. Also, it would be interesting to in- Heidelberg, 2005, vol. 3410, pp. 62–76.
vestigate whether the ranking results of DDR are consistent [22] D. Brockhoff, T. Wagner, and H. Trautmann, “On the Properties of
with those obtained by indicator-based ranking methods [64]. the R2 Indicator,” in Proceedings of the 14th Annual Conference on
Extending current work to solve constrained problems or Genetic and Evolutionary Computation, ser. GECCO ’12. Philadelphia,
Pennsylvania, USA: ACM, 2012, pp. 465–472.
dynamic problems is also one of further studies. [23] O. Schutze, X. Esquivel, A. Lara, and C. A. C. Coello, “Using the
Averaged Hausdorff Distance as a Performance Measure in Evolution-
ary Multiobjective Optimization,” IEEE Transactions on Evolutionary
R EFERENCES Computation, vol. 16, no. 4, pp. 504–522, Aug. 2012.
[24] R. Hernández Gómez and C. A. Coello Coello, “Improved Metaheuristic
[1] K. Miettinen, Nonlinear Multiobjective Optimization, ser. International
Based on the R2 Indicator for Many-Objective Optimization,” in Pro-
Series in Operations Research & Management Science, F. S. Hillier, Ed.
ceedings of the 2015 Annual Conference on Genetic and Evolutionary
Boston, MA: Springer US, 1998, vol. 12.
Computation, ser. GECCO ’15. Madrid, Spain: ACM, 2015, pp. 679–
[2] K. Deb, Multi-Objective Optimization Using Evolutionary Algorithms,
686.
1st ed., ser. Wiley-Interscience series in systems and optimization.
Chichester ; New York: John Wiley & Sons, 2001. [25] C. A. Rodrı́guez Villalobos and C. A. Coello Coello, “A new multi-
[3] J. D. Knowles and D. W. Corne, “Approximating the Nondominated objective evolutionary algorithm based on a performance assessment
Front Using the Pareto Archived Evolution Strategy,” Evolutionary indicator.” ACM Press, 2012, p. 505.
Computation, vol. 8, no. 2, pp. 149–172, Jun. 2000, 02038. [26] M. Li, S. Yang, and X. Liu, “Bi-goal evolution for many-objective
[4] D. W. Corne, N. R. Jerram, J. D. Knowles, M. J. Oates, and others, optimization problems,” Artificial Intelligence, vol. 228, pp. 45–65, Nov.
“PESA-II: Region-based selection in evolutionary multiobjective opti- 2015.
mization,” in Proceedings of the Genetic and Evolutionary Computation [27] B. Li, K. Tang, J. Li, and X. Yao, “Stochastic Ranking Algorithm
Conference (GECCO’2001), 2001. for Many-Objective Optimization Based on Multiple Indicators,” IEEE
[5] E. Zitzler, M. Laumanns, and L. Thiele, “SPEA2: Improving the strength Transactions on Evolutionary Computation, vol. 20, no. 6, pp. 924–938,
pareto evolutionary algorithm,” ETH Zurich, Tech. Rep. TIK-Report Dec. 2016.
103, 2001. [28] Y. Liu, D. Gong, J. Sun, and Y. Jin, “A Many-Objective Evolutionary
[6] K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, “A fast and elitist Algorithm Using A One-by-One Selection Strategy,” IEEE Transactions
multiobjective genetic algorithm: NSGA-II,” IEEE Transactions on on Cybernetics, pp. 1–14, 2017.
Evolutionary Computation, vol. 6, no. 2, pp. 182–197, Apr. 2002, 17703. [29] Qingfu Zhang and Hui Li, “MOEA/D: A Multiobjective Evolutionary
[7] B. Li, J. Li, K. Tang, and X. Yao, “Many-Objective Evolutionary Algorithm Based on Decomposition,” IEEE Transactions on Evolution-
Algorithms: A Survey,” ACM Computing Surveys, vol. 48, no. 1, pp. ary Computation, vol. 11, no. 6, pp. 712–731, Dec. 2007.
1–35, Sep. 2015. [30] K. Deb and H. Jain, “An Evolutionary Many-Objective Optimization Al-
[8] R. C. Purshouse and P. J. Fleming, “On the Evolutionary Optimization gorithm Using Reference-Point-Based Nondominated Sorting Approach,
of Many Conflicting Objectives,” IEEE Transactions on Evolutionary Part I: Solving Problems With Box Constraints,” IEEE Transactions on
Computation, vol. 11, no. 6, pp. 770–784, Dec. 2007. Evolutionary Computation, vol. 18, no. 4, pp. 577–601, Aug. 2014.
[9] M. Laumanns, L. Thiele, K. Deb, and E. Zitzler, “Combining Con- [31] M. Asafuddoula, T. Ray, and R. Sarker, “A Decomposition-Based Evolu-
vergence and Diversity in Evolutionary Multiobjective Optimization,” tionary Algorithm for Many Objective Optimization,” IEEE Transactions
Evolutionary Computation, vol. 10, no. 3, pp. 263–282, Sep. 2002, on Evolutionary Computation, vol. 19, no. 3, pp. 445–460, Jun. 2015.
00981. [32] Y. Yuan, H. Xu, B. Wang, B. Zhang, and X. Yao, “Balancing Conver-
[10] A. G. Hernández-Dı́az, L. V. Santana-Quintero, C. A. Coello Coello, and gence and Diversity in Decomposition-Based Many-Objective Optimiz-
J. Molina, “Pareto-adaptive -dominance,” Evolutionary Computation, ers,” IEEE Transactions on Evolutionary Computation, vol. 20, no. 2,
vol. 15, no. 4, pp. 493–517, Dec. 2007. pp. 180–198, Apr. 2016.
[11] Saku Kukkonen and Jouni Lampinen, “Ranking-Dominance and Many- [33] A. Trivedi, D. Srinivasan, K. Sanyal, and A. Ghosh, “A Survey of
Objective Optimization,” in IEEE Congress on Evolutionary Computa- Multiobjective Evolutionary Algorithms based on Decomposition,” IEEE
tion, 2007. IEEE, Sep. 2007, pp. 3983–3990. Transactions on Evolutionary Computation, vol. 21, no. 3, pp. 440–462,
[12] X. Zou, Y. Chen, M. Liu, and L. Kang, “A New Evolutionary Algorithm 2016.
for Solving Many-Objective Optimization Problems,” IEEE Transactions [34] K. Li, K. Deb, Q. Zhang, and S. Kwong, “An Evolutionary Many-
on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 38, no. 5, Objective Optimization Algorithm Based on Dominance and Decompo-
pp. 1402–1412, Oct. 2008. sition,” IEEE Transactions on Evolutionary Computation, vol. 19, no. 5,
[13] M. Li, S. Yang, and X. Liu, “Shift-Based Density Estimation for Pareto- pp. 694–716, Oct. 2015.
Based Algorithms in Many-Objective Optimization,” IEEE Transactions [35] R. Cheng, Y. Jin, M. Olhofer, and B. Sendhoff, “A Reference Vec-
on Evolutionary Computation, vol. 18, no. 3, pp. 348–365, Jun. 2014. tor Guided Evolutionary Algorithm for Many-Objective Optimization,”
[14] S. F. Adra and P. J. Fleming, “Diversity Management in Evolution- IEEE Transactions on Evolutionary Computation, vol. 20, no. 5, pp.
ary Many-Objective Optimization,” IEEE Transactions on Evolutionary 773–791, Oct. 2016.
Computation, vol. 15, no. 2, pp. 183–195, Apr. 2011. [36] Y. Yuan, H. Xu, B. Wang, and X. Yao, “A New Dominance Relation-
[15] Y. Xiang, Y. Zhou, M. Li, and Z. Chen, “A Vector Angle-Based Evo- Based Evolutionary Algorithm for Many-Objective Optimization,” IEEE
lutionary Algorithm for Unconstrained Many-Objective Optimization,” Transactions on Evolutionary Computation, vol. 20, no. 1, pp. 16–37,
IEEE Transactions on Evolutionary Computation, vol. 21, no. 1, pp. Feb. 2016.
131–152, Feb. 2017. [37] S. Jiang and S. Yang, “A Strength Pareto Evolutionary Algorithm Based
[16] X. Zhang, Y. Tian, and Y. Jin, “A Knee Point-Driven Evolutionary on Reference Direction for Multiobjective and Many-Objective Opti-
Algorithm for Many-Objective Optimization,” IEEE Transactions on mization,” IEEE Transactions on Evolutionary Computation, vol. 21,
Evolutionary Computation, vol. 19, no. 6, pp. 761–776, Dec. 2015. no. 3, pp. 329–346, Jun. 2017.
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. XX, SEPTEMBER 2017 15
[38] I. Das and J. E. Dennis, “Normal-Boundary Intersection: A New Computation (CEC), 2014 IEEE Congress On. IEEE, Jul. 2014, pp.
Method for Generating the Pareto Surface in Nonlinear Multicriteria 2156–2163.
Optimization Problems,” SIAM Journal on Optimization, vol. 8, no. 3, [60] E. Zitzler and L. Thiele, “Multiobjective evolutionary algorithms: A
pp. 631–657, Aug. 1998. comparative case study and the strength Pareto approach,” IEEE trans-
[39] H. Ishibuchi, Y. Setoguchi, H. Masuda, and Y. Nojima, “Performance of actions on Evolutionary Computation, vol. 3, no. 4, pp. 257–271, 1999.
Decomposition-Based Many-Objective Algorithms Strongly Depends on [61] D. A. Van Veldhuizen, “Multiobjective Evolutionary Algorithms: Classi-
Pareto Front Shapes,” IEEE Transactions on Evolutionary Computation, fications, Analyses, and New Innovations,” Ph.D. dissertation, Air Force
vol. 21, no. 2, pp. 169–190, Apr. 2017. Institute of Technology, Wright Patterson AFB, OH, USA, 1999.
[40] R. Saborido, A. B. Ruiz, and M. Luque, “Global WASF-GA: An [62] Y. Tian, R. Cheng, X. Zhang, and Y. Jin, “PlatEMO: A MATLAB
Evolutionary Algorithm in Multiobjective Optimization to Approximate Platform for Evolutionary Multi-Objective Optimization [Educational
the Whole Pareto Optimal Front,” Evolutionary Computation, pp. 1–41, Forum],” IEEE Computational Intelligence Magazine, vol. 12, no. 4,
Feb. 2016. pp. 73–87, Nov. 2017.
[41] S. Jiang and S. Yang, “An Improved Multiobjective Optimization Evolu- [63] N. Kowatari, A. Oyama, H. E. Aguirre, and K. Tanaka, “A Study
tionary Algorithm Based on Decomposition for Complex Pareto Fronts,” on Large Population MOEA Using Adaptive -Box Dominance and
IEEE Transactions on Cybernetics, vol. 46, no. 2, pp. 421–437, Feb. Neighborhood Recombination for Many-Objective Optimization,” in
2016. Learning and Intelligent Optimization. Berlin, Heidelberg: Springer
[42] Z. Wang, Q. Zhang, H. Li, H. Ishibuchi, and L. Jiao, “On the use of Berlin Heidelberg, 2012, vol. 7219, pp. 86–100.
two reference points in decomposition based multiobjective evolutionary [64] E. Zitzler and S. Künzli, “Indicator-Based Selection in Multiobjective
algorithms,” Swarm and Evolutionary Computation, vol. 34, pp. 89–102, Search,” in Parallel Problem Solving from Nature - PPSN VIII. Berlin,
Jun. 2017. Heidelberg: Springer Berlin Heidelberg, 2004, vol. 3242, pp. 832–842.
[43] X. Cai, Z. Mei, and Z. Fan, “A Decomposition-Based Many-Objective
Evolutionary Algorithm With Two Types of Adjustments for Direction Xiaoyu He received the B.Eng. degree from Beijing
Vectors,” IEEE Transactions on Cybernetics, pp. 1–14, 2017. Electronic Science and Technology Institute, Bei-
[44] S. Jiang, Z. Cai, J. Zhang, and Y.-S. Ong, “Multiobjective optimization jing, China, in 2010, and the M.P.A. degree from
by decomposition with Pareto-adaptive weight vectors,” in Seventh South China University of Technology, Guangzhou,
International Conference on Natural Computation, vol. 3. IEEE, Jul. China, in 2016. He is currently pursuing the Ph.D.
2011, pp. 1260–1264. degree with Sun Yat-sen University, Guangzhou.
[45] H. Jain and K. Deb, “An Evolutionary Many-Objective Optimization Al- His current research interests include evolutionary
gorithm Using Reference-Point Based Nondominated Sorting Approach, computation and data mining.
Part II: Handling Constraints and Extending to an Adaptive Approach,”
IEEE Transactions on Evolutionary Computation, vol. 18, no. 4, pp.
602–622, Aug. 2014.
[46] Y. Qi, X. Ma, F. Liu, L. Jiao, J. Sun, and J. Wu, “MOEA/D with Adaptive
Weight Adjustment,” Evolutionary Computation, vol. 22, no. 2, pp. 231– Yuren Zhou received the B.Sc. degree in mathemat-
264, Jun. 2014. ics from Peking University, Beijing, China, in 1988,
[47] M. Asafuddoula, H. K. Singh, and T. Ray, “An Enhanced the M.Sc. degree in Mathematics from Wuhan Uni-
Decomposition-Based Evolutionary Algorithm With Adaptive Reference versity, Wuhan, China, in 1991, and the Ph.D. degree
Vectors,” IEEE Transactions on Cybernetics, pp. 1–14, 2017. in Computer Science from the same University in
[48] H. L. Liu, L. Chen, Q. Zhang, and K. Deb, “Adaptively Allocating 2003. He is currently a professor with the School of
Search Effort in Challenging Many-Objective Optimization Problems,” Data and Computer Science, Sun Yat-sen University,
IEEE Transactions on Evolutionary Computation, vol. PP, no. 99, pp. Guangzhou, P. R. China.
1–1, 2017. His current research interests include design and
[49] C. A. R. Hoare, “Algorithm 65: Find,” Commun. ACM, vol. 4, no. 7, analysis of algorithms, evolutionary computation,
pp. 321–322, Jul. 1961. and social networks.
[50] D. P. Mitchell, “Spectrally optimal sampling for distribution ray tracing.”
ACM Press, 1991, pp. 157–164. Zefeng Chen received the B.Sc. degree in In-
[51] H.-L. Liu, F. Gu, and Q. Zhang, “Decomposition of a Multiobjective formation and Computational Science from Sun
Optimization Problem Into a Number of Simple Multiobjective Sub- Yat-sen University in 2013, and the M.Sc. degree
problems,” IEEE Transactions on Evolutionary Computation, vol. 18, in Computer Science and Technology from South
no. 3, pp. 450–455, Jun. 2014. China University of Technology in 2016. He is cur-
[52] H. Ishibuchi, K. Doi, and Y. Nojima, “On the effect of normalization in rently a Ph.D. candidate in Sun Yat-sen University,
MOEA/D for multi-objective and many-objective optimization,” Com- Guangzhou, P. R. China.
plex & Intelligent Systems, vol. 3, no. 4, pp. 279–294, Dec. 2017. His current research interests include evolution-
[53] H. K. Singh, A. Isaacs, and T. Ray, “A Pareto Corner Search Evolution- ary computation, multi-objective optimization, data
ary Algorithm and Dimensionality Reduction in Many-Objective Opti- mining, and machine learning.
mization Problems,” IEEE Transactions on Evolutionary Computation,
vol. 15, no. 4, pp. 539–556, Aug. 2011.
[54] M. Jensen, “Reducing the Run-Time Complexity of Multiobjective Qingfu Zhang (M’01-SM’06-F’17) received the
EAs: The NSGA-II and Other Algorithms,” IEEE Transactions on BSc degree in mathematics from Shanxi University,
Evolutionary Computation, vol. 7, no. 5, pp. 503–515, Oct. 2003. China in 1984, the MSc degree in applied mathemat-
[55] S. Kukkonen and J. Lampinen, “GDE3: The third Evolution Step ics and the PhD degree in information engineering
of Generalized Differential Evolution,” in 2005 IEEE Congress on from Xidian University, China, in 1991 and 1994,
Evolutionary Computation, vol. 1, Sep. 2005, pp. 443–450 Vol.1. respectively.
[56] S. Huband, P. Hingston, L. Barone, and L. While, “A review of He is a Professor at the Department of Com-
multiobjective test problems and a scalable test problem toolkit,” IEEE puter Science, City University of Hong Kong, Hong
Transactions on Evolutionary Computation, vol. 10, no. 5, pp. 477–506, Kong. His main research interests include evolu-
Oct. 2006. tionary computation, optimization, neural networks,
[57] K. Deb, L. Thiele, M. Laumanns, and E. Zitzler, “Scalable Test Problems data analysis, and their applications. He is currently
for Evolutionary Multiobjective Optimization,” in Evolutionary Multi- leading the Metaheuristic Optimization Research (MOP) Group in City
objective Optimization, A. Abraham, L. Jain, and R. Goldberg, Eds. University of Hong Kong.
London: Springer-Verlag, 2005, pp. 105–145. Dr. Zhang is an Associate Editor of the IEEE Transactions on Evolutionary
[58] R. Cheng, M. Li, Y. Tian, X. Zhang, S. Yang, Y. Jin, and X. Yao, Computation and the IEEE Transactions on Systems, Man, and Cybernetics-
“A benchmark test suite for evolutionary many-objective optimization,” Part B. He is also an Editorial Board Member of three other international
Complex & Intelligent Systems, vol. 3, no. 1, pp. 67–81, Mar. 2017. journals. He is a Web of Science highly cited researcher in Computer Science.
[59] H. Li, Q. Zhang, and J. Deng, “Multiobjective test problems with
complicated Pareto fronts: Difficulties in degeneracy,” in Evolutionary