Elsevier

Applied Soft Computing

Volume 12, Issue 11, November 2012, Pages 3526-3538
Applied Soft Computing

A regularity model-based multiobjective estimation of distribution algorithm with reducing redundant cluster operator

https://doi.org/10.1016/j.asoc.2012.06.008Get rights and content

Abstract

A regularity model-based multiobjective estimation of distribution algorithm (RM-MEDA) has been proposed for solving continuous multiobjective optimization problems with variable linkages. RM-MEDA is a kind of estimation of distribution algorithms and, therefore, modeling plays a critical role. In RM-MEDA, the population is split into several clusters to build the model. Moreover, the fixed number of clusters is recommended in RM-MEDA when solving different kinds of problems. However, based on our experiments, we find that the number of clusters is problem-dependent and has a significant effect on the performance of RM-MEDA. Motivated by the above observation, in this paper we improve the clustering process and propose a reducing redundant cluster operator (RRCO) to build more precise model during the evolution. By combining RRCO with RM-MEDA, we present an improved version of RM-MEDA, named IRM-MEDA. In this paper, we also construct four additional continuous multiobjective optimization test instances. The experimental results have shown that IRM-MEDA outperforms RM-MEDA in terms of efficiency and effectiveness. In particular, IRM-MEDA performs on average 31.67% faster than RM-MEDA.

Highlights

► Analyze the drawback of modeling in RM-MEDA. ► Propose a reducing redundant clustering operator (RRCO). ► By combining RRCO with RM-MEDA, propose an improved version of RM-MEDA, named IRM-MEDA. ► Verify the effectiveness of RRCO by many experiments.

Introduction

Many optimization problems involve not one but several objectives which should be optimized simultaneously. This kind of problems is considered as multiobjective optimization problems (MOPs). In this paper, we consider the following continuous MOPs:minimizey=f(x)=(f1(x),f2(x),,fm(x))where x=(x1,,xn)XRn is the decision vector, X is the decision space, yYRm is the objective vector, and Y is the objective space.

There are some basic definitions in multiobjective optimization, which are introduced as follows.

Definition 1

Given two decision vectors a=(a1,,an) and b=(b1,,bn), if ∀i  {1, …, m} , fi(a)fi(b) and ∃j  {1, …, m} , fj(a)<fj(b), we say a Pareto dominates b, denoted as ab.

Definition 2

A decision vector xX is called Pareto optimal solution if there does not exist another decision vector xX such that xx.

Definition 3

The Pareto set (PS) is the set of all the Pareto optimal solutions:PS={xX|¬xX,xx}The solutions in the PS are also called nondominated solutions.

Definition 4

The Pareto front (PF) is the set of the objective vectors of all the Pareto optimal solutions:PF={f(x)|xPS}

For MOPs, in most cases, we cannot find a single solution to optimize all the objectives at the same time. Therefore, we have to balance them and find a set of optimal tradeoffs, i.e., Pareto set (PS) in the decision space and Pareto front (PF) in the objective space, respectively. Since evolutionary algorithms (EAs) deal with a group of candidate solutions simultaneously, it seems to be natural to use EAs for finding a group of Pareto optimal solutions when solving MOPs. Vector evaluation genetic algorithm (VEGA), introduced by Schaffer [1] in 1980s, is the first actual implementation of EAs to solve MOPs. After that, a considerable number of multiobjective evolutionary algorithms (MOEAs) have been proposed due to increasing interest in solving MOPs by EAs.

The development of MOEAs can be briefly divided into three generations [2], [3]. In the first generation of MOEAs, Pareto ranking and fitness sharing are the most common techniques adopted by MOEAs. There are some paradigms in this generation, for example: nondominated sorting genetic algorithm (NSGA), proposed by Srinivas and Deb [4], is based on several layers of classifications of the individuals as suggested by Goldberg [5] and uses crowding distance to maintain the diversity of the population. Niched-Pareto genetic algorithm (NPGA), proposed by Horn et al. [6], employs tournament selection based on Pareto dominance and fitness sharing to keep the diversity. Fonseca and Fleming [7] introduced a multiobjective genetic algorithm (MOGA).

The second generation of MOEAs is characterized by the elitism preservation, which usually stores the nondominated individuals into a predefined archive (also called external population). It is necessary to note that incorporating the elitism into MOEAs can facilitate the convergence of the population. Zitzler and Thiele [8] proposed strength Pareto EA (SPEA), which uses an archive to store the nondominated solutions found so far and adopts clustering to prune the archive if the number of nondominated individuals in the archive exceeds a predefined value. Zitzler et al. [9] also proposed an improved version of SPEA, referred as SPEA2. Compared with SPEA, SPEA2 has the following three properties: (1) a new fitness assignment strategy, (2) a density estimation technique, and (3) a novel archive truncation method. Knowles and Corne [10] presented Pareto archive evolutionary strategy (PAES), which uses (1 + 1)-ES to generate offspring. In PAES, the offspring is compared with the parent and the previously archived nondominated individuals for survival. Moreover, PAES divides the objective space into grids, the aim of which is to maintain the diversity of the population. Inspired by PAES, Corne et al. further developed PESA [11] and PESA-II [12]. Deb et al. [13] proposed an improved version of NSGA, called NSGA-II, by incorporating a fast nondominated sorting approach and a crowding-comparison approach.

In the current research, which belongs to the third generation of MOEAs, some new dominance concepts other than traditional Pareto dominance have been introduced. For instance, Laumanns et al. [14] introduced ɛ-dominance. Hernández-Díaz et al. [15] proposed an adaptive ɛ-dominance, which is an improvement of the original ɛ-dominance [14]. Ben Said et al. [16] proposed r-dominance for interactive evolutionary multi-criteria decision making. Brockoff and Zitzler [17] proposed a local dominance scheme to reduce objective dimensionality. In addition, some researchers combined traditional weight vector based techniques with EAs to deal with MOPs [18], [19], [20], [21]. Recently, Zhang and Li [22] proposed a novel MOEA based on decomposition, called MOEA/D, which converts MOPs into a set of scalar optimization subproblems. Moreover, MOEA/D utilizes the neighbor information to produce offspring and optimize the subproblems simultaneously.

Many attempts have also been made to improve the performance of MOEAs by making use of different kinds of EAs as well as swarm intelligence. For example, Coello Coello et al. [23] incorporated Pareto dominance into particle swarm optimization for solving MOPs. Li and Zhang [24] proposed a new version of MOEA/D [22] based on differential evolution. Igel et al. [25] developed a variant of covariance matrix adaptation evolution strategy (CMA-ES) [26] for multiobjective optimization. Ghoseiria and Nadjari [27] presented an algorithm based on multiobjective ant colony optimization to solve the bi-objective shortest path problem. Jamuna and Swarup [28] proposed a multiobjective biogeography based optimization algorithm to design optimal placement of phasor measurement units. Zhang [29] proposed an immune optimization algorithm for dealing with constrained nonlinear multiobjective optimization problems.

Recently, indicator-based MOEAs have also been actively researched in the community of evolutionary multiobjective optimization [30], [31].

It can be induced from the Karush–Kuhn–Tucker condition that the PS of a continuous MOP is a (m−1)-dimensional piecewise continuous manifold in the decision space [32], [33], where m is the number of objectives. Thus, for the continuous biobjective optimization problems (i.e., m = 2), the PS is a piecewise continuous curve; and for the continuous triobjective optimization problems (i.e., m = 3), the PS is a piecewise continuous 2-D surface.

Based on the above regularity, Zhang et al. [34] proposed a regularity model-based multiobjective estimation of distribution algorithm, referred as RM-MEDA. As a kind of estimation of distribution algorithms (EDAs) [35], RM-MEDA employs the (m−1)-dimensional local principal component analysis ((m−1)-D local PCA) [36] to build the model of the PS in the decision space. The (m−1)-D local PCA is a locally linear approach to nonlinear dimension reduction, which can construct local models, each pertaining to a different disjoint region of the data space. In RM-MEDA, firstly, the (m−1)-D local PCA divides the population into K (K is a constant integer) disjoint clusters and computes the central point and principal component of each cluster. Afterward, one model is built based on the corresponding central point and principal component for each cluster. The primary aim of modeling in RM-MEDA is to approximate one of the pieces of the PS by making use of the solutions in one cluster. Ideally, if the number of clusters K is equal to the number of the pieces of the PS, each piece of the PS can be approximated by one cluster. In this case, a precise model may be built and the performance of RM-MEDA may be excellent. However, if the number of clusters K is not equal to the number of the pieces of the PS; needless to say, the model is not precise.

Since we have no priori knowledge about the number of the pieces of the PS for a MOP at hand, it is very difficult to determine a reasonable value for K. Moreover, the setting of K is usually problem-dependent. In particular, based on our experiments, this parameter has a significant effect on the performance of RM-MEDA. Since K is fixed to 5 in RM-MEDA, this setting might not be very effective for different kinds of MOPs. In order to overcome the above drawback of RM-MEDA, we design a reducing redundant cluster operator (RRCO) to enhance the modeling precision of RM-MDEA. By integrating RRCO with RM-MEDA, IRM-MEDA is derived. Extensive experiments have been conducted to compare IRM-MEDA with its predecessor RM-MEDA on a set of biobjective and triobjective test instances with variable linkages (note that variable linkages reflect the interactions among the variables). The experimental results verify that the efficiency and effectiveness of RM-MEDA can be significantly improved by RRCO.

The rest of the paper is organized as follows. Section 2 briefly reviews RM-MEDA. The drawback of modeling in RM-MEDA is discussed in Section 3. Section 4 presents the details of RRCO. IRM-MEDA is described in Section 5. The experimental results are reported in Section 6. Finally, Section 7 concludes this paper.

Section snippets

Framework

During the evolution, RM-MEDA maintains:

  • a population Pt of N individuals: Pt={x1,,xN}, where t is the generation number;

  • their f-values:f(x1),,f(xN).

RM-MEDA is implemented as follows:

Step 1Initialization: Generate an initial population P0 by randomly sampling N individuals from the decision space S and compute the f-values of these individuals.
Step 2Modeling: According to the population Pt, build the probability model by the (m−1)-D local PCA.
Step 3Sampling: Generate an offspring

The drawback of modeling in RM-MEDA

According to the introduction in Section 2, we can see that modeling is a very important process in RM-MEDA. Moreover, more promising solutions may be generated by sampling from a more precise model. Since the PS of the continuous MOPs is a piecewise continuous (m−1)-dimensional manifold, RM-MEDA firstly partitions the population Pt into K clusters C1, …, CK by making use of the (m−1)-D local PCA, with the aim of estimating one manifold ψi with one cluster Ci. However, a question which

Reducing redundant cluster operator

Although RM-MEDA usually sets the number of clusters with a big value, as analyzed previously, a big value will also have a side effect on the performance. Moreover, the setting of the number of clusters is usually problem-dependent. In order to overcome the above drawback of RM-MEDA, we propose a reducing redundant cluster operator (RRCO) in this paper. In the following, the details of RRCO are introduced.

According to the discussion in Section 3, if there exist some redundant clusters, two

IRM-MEDA

By combining RRCO with RM-MEDA, we present an improved version of RM-MEDA, referred as IRM-MEDA. The framework of IRM-MEDA is the same as that of RM-MEDA except that IRM-MEDA uses RRCO to update K. In RM-MEDA, K is fixed during the evolution, whereas IRM-MEDA dynamically adjusts K by extracting some information from the result of clustering at each generation.

IRM-MEDA performs as follows:

Step 1Initialization: Randomly generate an initial population P0={x1,x2,,xN} and initialize K (i.e., the

Test instances

In this paper, we use ten test instances (F1–F10) to verify the effectiveness of IRM-MEDA. The main information of these ten test instances is summarized in Table 1. Test instances F1–F6 are taken from [34]. Since the primary purpose of RM-MEDA is to approximate the PS of MOPs by taking advantage of the (m−1)-D local PCA, we construct four additional test instances (i.e., F7–F10) with various PS structures to further compare RM-MEDA with IRM-MEDA. It is necessary to note that the PS structures

Conclusion

RM-MEDA [34] is a recently proposed approach for solving MOPs with variable linkages. In order to approximate the PS of MOPs, the (m−1)-D local PCA is adopted to build the model of the population in RM-MEDA. In this paper, we analyze the drawback of the clustering process in the modeling suggested by RM-MEDA. Based on our analysis, the number of clusters is problem-dependent and has a significant effect on the performance of RM-MEDA. However, a fixed number of clusters is recommended in [34]

Acknowledgments

The authors sincerely thank the anonymous reviewers for their constructive and helpful comments and suggestions.

This research was supported in part by the National Natural Science Foundation of China under Grant 60805027, 61175064 and 90820302, and in part by the Research Fund for the Doctoral Program of Higher Education under Grant 200805330005.

References (43)

  • Z. Zhang

    Immune optimization algorithm for constrained nonlinear multiobjective optimization problems

    Applied Soft Computing

    (2007)
  • J.D. Schaffer

    Multiple objective optimization with vector evaluated genetic algorithms

    Proceedings of the First International Conference on Genetic Algorithms

    (1985)
  • C.A. Coello Coello

    Evolutionary multi-objective optimization: a historical view of the field

    IEEE Computational Intelligence Magazine

    (2006)
  • M. Gong et al.

    Research on evolutionary multi-objective optimization algorithms

    Journal of Software

    (2009)
  • N. Srinivas et al.

    Multiobjective optimization using nondominated sorting in genetic algorithms

    Evolutionary Computation

    (1994)
  • D.E. Goldberg

    Genetic Algorithms in Search, Optimization and Machine Learning

    (1989)
  • J. Horn et al.

    A Niched Pareto genetic algorithm for multiobjective optimization

    Proceedings of the First IEEE Conference on Evolutionary Computation, IEEE World Congress on Computational Intelligence

    (1994)
  • C.M. Fonseca et al.

    Genetic algorithms for multiobjective optimization: formulation, discussion and generalization

    Proceedings of the Fifth International Conference on Genetic Algorithms

    (1993)
  • E. Zitzler et al.

    Multiobjective evolutionary algorithms: a comparative case study and the strength Pareto approach

    IEEE Transactions on Evolutionary Computation

    (1999)
  • E. Zitzler et al.

    SPEA2: improving the strength Pareto evolutionary algorithm for multiobjective optimization

  • J.D. Knowles et al.

    Approximating the nondominated front using the Pareto archived evolution strategy

    Evolutionary Computation

    (2000)
  • D.W. Corne et al.

    The Pareto envelope-based selection algorithm for multiobjective optimization

  • D.W. Corne et al.

    PESA-II: region-based selection in evolutionary multiobjective optimization

  • K. Deb et al.

    A fast and elitist multiobjective genetic algorithm: NSGA-II

    IEEE Transactions on Evolutionary Computation

    (2002)
  • M. Laumanns et al.

    Combining convergence and diversity in evolutionary multi-objective optimization

    Evolutionary Computation

    (2002)
  • A.G. Hernández-Díaz et al.

    Pareto-adaptive epsilon-dominance

    Evolutionary Computation

    (2007)
  • L. Ben Said et al.

    The r-dominance: a new dominance relation for interactive evolutionary multicriteria decision making

    IEEE Transactions on Evolutionary Computation

    (2010)
  • D. Brockoff et al.

    Are all objective necessary on dimensionality reduction in evolutionary multi-objective optimization

    Proceeding of Parallel Problem Solving from Nature, PPSN IX, LNCS

    (2006)
  • H. Ishibuchi et al.

    Multi-objective genetic local search algorithm and its application to flowshop scheduling

    IEEE Transactions on Systems, Man and Cybernetics

    (1998)
  • H. Ishibuchi et al.

    Balance between genetic search and local search in memetic algorithm for multiobjective permutation flowshop scheduling

    IEEE Transactions on Evolutionary Computation

    (2003)
  • Y.W. Leung et al.

    Multiobjective programming using uniform design and genetic algorithm

    IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews

    (2000)
  • Cited by (43)

    • A practical regularity model based evolutionary algorithm for multiobjective optimization

      2022, Applied Soft Computing
      Citation Excerpt :

      The regularity model based approaches have been well studied and applied in EMO. In [30], a regularity model with a reducing redundant cluster operator is proposed to build more precise models during the evolution. As suggested in [31], the regularity model is embedded into existing MOEAs to deal with noisy MOPs.

    • A dynamic multi-objective particle swarm optimization algorithm based on adversarial decomposition and neighborhood evolution

      2022, Swarm and Evolutionary Computation
      Citation Excerpt :

      The re-initialized population includes three parts: prediction solution set, non-dominated solution set, and random initialization solution set. Zhou et al. [47] developed a population prediction strategy (PPS) based on a regularity model-based multiobjective estimation of distribution algorithm (RM-MEDA) [48] to solve DMOPs. In PPS, the population is divided into population centers and manifolds, and an autoregressive time series forecasting technique is applied to guide the estimation of new locations.

    View all citing articles on Scopus
    View full text