Elsevier

Information Sciences

Volumes 334–335, 20 March 2016, Pages 44-64
Information Sciences

Biasing the transition of Bayesian optimization algorithm between Markov chain states in dynamic environments

https://doi.org/10.1016/j.ins.2015.11.030Get rights and content

Abstract

When memory-based evolutionary algorithms are applied in dynamic environments, the certainly use of uncertain prior knowledge for future environments may mislead the evolutionary algorithms. To address this problem, this paper presents a new, memory-based evolutionary approach for applying the Bayesian optimization algorithm (BOA) in dynamic environments. Our proposed method, unlike existing memory-based methods, uses the knowledge of former environments probabilistically in future environments. For this purpose, the run of BOA is modeled as the movements in a Markov chain, in which the states become the Bayesian networks that are learned in every generation. When the environment changes, a stationary distribution of the Markov chain is defined on the basis of the retrieved prior knowledge. Then, the transition probabilities of BOA in the Markov chain are modified (biased) to comply with the defined stationary distribution. To this end, we employ the Metropolis algorithm and modify the K2 algorithm for learning the Bayesian network in BOA in order to reflect the obtained transition probabilities. Experimental results show that the proposed method achieves improved performance compared to conventional methods, especially in random environments.

Introduction

For many real-world problems, their environments might change during the optimization process. Environmental changes can take various forms such as changes in the parameters, objective functions, or constraints [3], [31], [50]. When the environment varies, the optimum of that problem tends to be altered as well. Dynamic optimization problems may take into account any of these uncertain events in the optimization process [18], [30], [46]. Examples include scheduling problems where new jobs arrive over time [4], [43], vehicle routing problems where new request arrivals vary with time [4], portfolio-decision optimization problems where the stock market changes and the decisions should be updated consistently [3], and flight path optimization problems where the flight paths should be adapted to special events [3]. In the dynamic optimization problems, the objective is not only finding the optimum solution but also tracking it in a timely manner [51]. In cases where the problem after the environmental change remains similar to previous problems, we can use the knowledge acquired while searching for previous solutions to find the solution to the new problem [18]. Thus, searching for a new solution based on the previous solutions can save computation time [45].

Many real-world dynamic problems have been solved by various soft computing techniques, such as neural networks, particle swarm optimization, and ant colony algorithm [5], [6], [24], [26], [33], [44], [47], [54]. Evolutionary algorithms inspired by natural evolution can be especially powerful tools for solving dynamic optimization problems [51]. As a class of evolutionary algorithms [10], estimation of distribution algorithms (EDAs) [22], [36], [37] have been proposed to resolve the “building blocks destruction” problem. In EDAs, instead of using the standard (genetic) operators, probabilistic models that describe high-quality solutions are used to generate new solutions [29]. Various versions of EDAs differ in their probabilistic models. Meanwhile, numerous evolutionary methods have been proposed to handle the dynamic optimization problems. These methods are classified as follows [2]: (a) reaction of the algorithm after the change, (b) preserving the diversity during the run, (c) memory-based approach, and (d) multi-population approach. In the memory-based approach, evolutionary algorithms are supported by a memory that stores useful information regarding the optimization problems in the previous environments. Compared to other evolutionary algorithms, EDAs have more potential to be used as memory-based dynamic optimization approaches because their probabilistic models contain abstract information about the previous environments [38]. These models can be efficiently stored in the memory, and this memory can be managed in a simple, rapid manner [38]. When environmental change occurs, the relevant information is retrieved and used to solve the new problem [18].

As mentioned, EDAs have greater potential than other evolutionary algorithms to be used for memory-based dynamic optimization approach, and some researchers have successfully applied EDAs to memory-based dynamic optimization. Section 2 reviews these studies.

There is a common feature in all EDA approaches: knowledge of the previous environments is stored in the memory; when a change occurs, the relevant knowledge is retrieved and utilized directly to solve the new problem. When the environmental changes are repetitive or the new environments are similar to the former ones, the retrieved knowledge can be compatible with the new environments. Hence, memory-based methods that deterministically utilize former knowledge are highly efficient. When a change occurs in new environments that are not sufficiently similar to the former ones, no compatible knowledge will be found in the memory. In addition, the similarity measure used for searching the memory and finding the relevant knowledge might not be reliable. In that case, knowledge that seems compatible in the new environment would mislead evolutionary algorithms, and making use of the retrieved knowledge in a deterministic manner would not be effective. The performance achieved with this knowledge may even be lower than when no prior knowledge is used because this inaccurate knowledge can bias evolutionary algorithms toward an incorrectly estimated region in the search space, resulting in the algorithms becoming stuck at sub-optimal points.

In this paper, we address the problem of memory-based evolutionary algorithm misguidance from use of inaccurate prior knowledge. We study the probabilistic use of uncertain prior knowledge in dynamic environments and develop, using Markov chain modeling, a new memory-based method that employs the BOA. To this end, the run of BOA in the dynamic environments is modeled as the movements between the states in a heterogeneous Markov chain. When the environment changes, relevant knowledge is retrieved from the memory and it is used to define a stationary distribution (which persists forever once it is reached) for visiting the states of the Markov chain. Then the transitions of BOA between the states are controlled in order to satisfy the defined stationary distribution.

In our proposed method, some stages are inspired from our previously presented method called DBN-MBOA [19] in order to improve the method. DBN-MBOA [19] is a memory-based BOA that uses prior knowledge certainly and, therefore, like other memory-based evolutionary algorithms, suffers from the misguidance problem. Here, we attempt to resolve that problem by using the prior knowledge probabilistically.

The rest of the paper is organized as follows. Section 2 reviews the previous studies conducted on memory-based dynamic optimization. Section 3 briefly reviews the BOA and the DBN-MBOA. In Section 4, the stochastic process of “transition of BOA between consecutive generations” is modeled as “the movements among the states of a heterogeneous Markov chain”; thus, we control the movement of BOA among the states of Markov chain to address optimization in dynamic environments. Section 5 describes in detail the new method to bias the transition of BOA in the Markov chain along with its mathematical relations. The section also discusses the characteristics of the Markov chain in the proposed algorithm. The experimental results are demonstrated in Section 6; finally, Section 7 concludes the paper with a summary.

Section snippets

Literature review

This section reviews the previous studies on memory-based dynamic optimization. The first work, univariate marginal distribution algorithm [29], the simplest version of EDAs, was applied to dynamic optimization problems [49]. In [49], the best samples created by the probability vector of univariate marginal distribution algorithm are stored in the memory, which is re-evaluated in every generation. If the fitness of any memory element varies, the environment is detected to be changed. The memory

Background

This section presents the background knowledge of the study. The BOA is explained in Sections 3.1, and 3.2 reviews the DBN-MBOA that provides the prior knowledge in our work.

Modeling BOA's evolutionary process as a Markov chain

Several mathematical models such as the Markov model [17], [40], Hollands’ schema theorem [16], and drift analysis [14], [15] have been used to study and model the behavior of evolutionary algorithms (e.g., EDAs and genetic algorithms). Among these, the Markov model is a powerful and common model that allows a study of the behavior of evolutionary algorithms over their evolution trajectories. Therefore, the Markov model is suitable for the present study in which we attempted to manage the

Proposed approach: BT-MBOA

In Section 4, the evolutionary process of BOA was modeled as a Markov chain. In general, the evolution direction in BOA is affected by two forces [9]. The first force is natural selection [9]. According to the natural selection rule, in each generation of BOA, the candidate solutions with the highest fitness are selected and used to create the Bayesian networks; thus, the probabilities of visiting the states in this Markov chain (i.e., the probabilities of creating the Bayesian networks) depend

Evaluation and results

In this section, we compare the results of BT-MBOA with those of EI-MBOA [38] and DBN-MBOA [19], which are among the most recent methods to apply the BOA in dynamic environments. The DBN-MBOA and the EI-MBOA are memory-based approaches that utilize the prior knowledge deterministically for the new environments. These methods were briefly reviewed in Section 3. In this experiment, we used a dynamic problem generator to construct various dynamic environments on the basis of the stationary

Conclusions

In this paper, the transition of BOA from one generation to another was modeled as the transition in a Markov chain in which the states represent the Bayesian networks learned in every generation. To apply the BOA in dynamic environments, the knowledge obtained from the previous environments was used to define a promising stationary distribution for this Markov chain. Then, we modified the K2 algorithm for learning the Bayesian network (based on the BIC metric) so that the transition of BOA

Acknowledgment

This paper was supported by Samsung Research Fund, Sungkyunkwan University, 2015.

References (54)

  • R. Tinós et al.

    Analysis of fitness landscape modifications in evolutionary dynamic optimization

    Inf. Sci. (Ny)

    (2014)
  • A.M. Turky et al.

    A multi-population harmony search algorithm with external archive for dynamic optimization problems

    Inf. Sci. (Ny)

    (2014)
  • S. Baluja, Population-based incremental learning: A method for integrating genetic search based function optimization...
  • J. Branke

    Evolutionary Optimization in Dynamic Environments

    (2002)
  • J. Branke et al.

    Multiobjective optimization for dynamic environments

  • J. Branke et al.

    Towards an analysis of dynamic environments

  • G.F. Cooper et al.

    A Bayesian method for the induction of probabilistic networks from data

    Mach. Learn.

    (1992)
  • T.E. Davis

    Toward an Extrapolation Of The Simulated Annealing Convergence Theory Onto The Simple Genetic Algorithm

    (1991)
  • G. Dick et al.

    The behaviour of genetic drift in a spatially-structured evolutionary algorithm

  • D.E. Goldberg

    Genetic Algorithms in Search, Optimization and Machine Learning

    (1989)
  • C. Gonzalez

    Contributions on Theoretical Aspects of Estimation of Distribution Algorithms

    (2005)
  • C. Gonzalez et al.

    Analysis of the univariate marginal distribution algorithm modeled by Markov chains

  • W.K. Hastings

    Monte Carlo sampling methods using Markov chains and their applications

    Biometrika.

    (1970)
  • J. He et al.

    A study of drift analysis for estimating computation time of evolutionary algorithms

    Nat. Comput.

    (2004)
  • J.H. Holland

    Adaptation in Natural and Artificial Systems

    (1975)
  • Y. Hong et al.

    Convergence of estimation of distribution algorithms in optimization of additively noisy fitness functions

  • Y. Jin et al.

    Evolutionary optimization in uncertain environments—a survey

    IEEE Trans. Evol. Comput.

    (2005)
  • Cited by (8)

    • Experimental analysis of a statistical multiploid genetic algorithm for dynamic environments

      2022, Engineering Science and Technology, an International Journal
      Citation Excerpt :

      In the work, they put immigrants into the population after a change instead of putting every generation. A BOA-based EA for dynamic environments is proposed in [21]. The proposed method starts with a randomly formed population.

    • Simultaneous optimization of cluster head selection and inter-cluster routing in wireless sensor networks using a 2-level genetic algorithm

      2022, Applied Soft Computing
      Citation Excerpt :

      For this purpose, the route robustness can be considered as an additional criterion in finding the optimal routes and the robust genetic algorithm [55,56] can be applied in the second stage of the proposed method instead of standard genetic algorithm. Another suggestion is to apply the dynamic genetic algorithm [57–59] in the second stage of the proposed method to adaptively re-adjust the route whenever any of the cluster heads dies. In addition, it is recommended that the future studies apply methods other than evolutionary algorithms (e.g., learning automata [60]) to solve the two problems in the same phase and evaluate the results.

    • Robust optimization using Bayesian optimization algorithm: Early detection of non-robust solutions

      2017, Applied Soft Computing Journal
      Citation Excerpt :

      Moreover, these features on DUF4 are presented in Table 4. As seen in Tables 3 and 4, the population sizes are equal to the problem sizes, by referring to former studies [35–38]. For simplicity in a comparative study, the same population size is used for all the three algorithms.

    View all citing articles on Scopus
    View full text