Performance analysis of dynamic optimization algorithms using relative error distance

https://doi.org/10.1016/j.swevo.2021.100930Get rights and content

Abstract

Quantification of the performance of algorithms that solve dynamic optimization problems (DOPs) is challenging, since the fitness landscape changes over time. Popular performance measures for DOPs do not adequately account for ongoing fitness landscape scale changes, and often yield a confounded view of performance. Similarly, most popular measures do not allow for fair performance comparisons across multiple instances of the same problem type nor across different types of problems, since performance values are not normalized. Many measures also assume normally distributed input data values, while in reality the necessary conditions for data normality are often not satisfied. The majority of measures also fail to capture the notion of performance variance over time. This paper proposes a new performance measure for DOPs, namely the relative error distance. The measure shows how close to optimal an algorithm performs by considering the multi-dimensional distance between the vector comprising the normalized performance scores for specific algorithm iterations of interest, and the theoretical point of best possible performance. The new measure does not assume normally distributed performance data across fitness landscape changes, is resilient against fitness landscape scale changes, better incorporates performance variance across fitness landscape changes into a single scalar value, and allows easier algorithm comparisons using established nonparametric statistical methods.

Introduction

Dynamic optimization problems (DOPs) are challenging since, as time passes, optimal solutions may become sub-optimal while previously inferior regions of the search space may suddenly yield the best solutions. Most optimization algorithms that are specialized to solve problems in static environments tend to be ineffective in dynamic environments due to the dynamic nature of DOPs [1]. As a result, many swarm intelligence (SI) and evolutionary computation (EC) meta-heuristics have been developed that focus specifically on solving DOPs [2], [3], [4], [5].

Quantification of the performance of algorithms is not as straightforward for dynamic environments as it is for static environments. Cruz et al. [2] outline how, for static environments, it often suffices to report just the quality of the best-found solution, and (optionally) the memory/time cost to achieve the result. For DOPs, these simple assessments are not sufficient since practitioners are interested in other aspects of algorithm performance during the optimization process. Such details include understanding an algorithm’s ability to detect problem changes, continually explore and exploit different areas of the problem search space, track existing optima as they move through the problem space, find new optima as they appear over time, track the stability of solutions over time, and determine how constraints are handled.

Researchers have defined many specialized performance measures to assess algorithms that solve DOPs. A number of surveys discuss the most popular performance measures that are used to quantify the ability of algorithms in solving DOPs [2], [4], [5]. The majority of reviewed measures have shortcomings that make fair comparisons of algorithm performance challenging. These shortcomings are discussed at length in this paper. The most often-used measures in literature, namely the offline error and offline performance [6], [7], and average best error before change [8] have crippling limitations that can cause observers to misreport findings by up to 60% (as shown in Section 5). Such shortcomings make it hard for existing measures to faithfully represent the underlying reality of an algorithm’s performance while solving a DOP. This paper proposes a new performance measure called the relative error distance (RED), that addresses the limitations of popular performance measures. The paper therefore makes a contribution towards more fair and sound measures to analyze and compare the performance of algorithms focused on solving DOPs.

The paper outline is as follows. Section 2 provides background information on DOPs, performance measures for algorithms aimed at solving DOPs, and a discussion of the attributes and shortcomings of popular performance measures. Section 3 introduces the RED measure. Section 4 outlines the experimental approach to validate the noteworthy characteristics of the RED measure, and section 5 presents the results of the empirical investigation. Section 6 concludes the paper.

Section snippets

Dynamic optimization problems

A working definition of a DOP is provided below, followed by an outline of commonly used DOP-focused performance measures, as well as their attributes and shortcomings.

Relative error distance

This paper proposes a new performance measure called relative error distance (RED) that helps to address the shortcomings of existing performance measures. Consider the vector b=(b1,b2,...,bnv) which represents nv measured performance values of a single execution (or run) of an algorithm. The components of b, namely bi, consist of nv different RE values (as defined in either Eqs. (10) or (11)). The exact nv algorithm iterations that are considered by the RED measure is configurable. That is, if

Empirical validation of relative error distance

The experiments in this study investigate whether the characteristics observed in actual algorithm error/fitness value data warrant the need for the proposed RED measure. The purpose is to obtain a reasonably large and diverse set of performance scores over many different types of problems. The performance comparison of specific algorithms against state-of-the-art methods is not the goal of the analysis.

The following four questions are answered:

  • Normally distributed error/fitness data: Are the

Results

The results of the experiments laid out in Section 4 are presented below.

Conclusion

The dynamic time-dependent nature of dynamic optimization problems (DOPs) makes it complex to objectively capture the performance of algorithms. Empirical investigations in this paper highlighted that, overwhelmingly, the series of performance values yielded by computational intelligence algorithms while solving a DOP is likely to contain significant fitness scale changes over time, is unlikely to follow a Gaussian distribution, and is likely to show significant variance over time. Popular

CRediT authorship contribution statement

Stéfan A.G. van der Stockt: Conceptualization, Formal analysis, Investigation, Methodology, Software, Writing – original draft, Writing – review & editing. Gary Pamparà: Conceptualization, Methodology, Software, Writing – review & editing. Andries P. Engelbrecht: Supervision, Writing – review & editing. Christopher W. Cleghorn: Supervision, Writing – review & editing.

Declaration of Competing Interest

We the authors, Stefan van der Stockt, Gary Pamparà, Andries Engelbrecht, and Christopher Cleghorn confirm that there are no conflicts of interest with any of the authors as it pertains to publishing this work. This work is the result of Stefan and Gary’s PhD research that is getting submitted for publication.

References (70)

  • C. Cruz et al.

    Optimization in dynamic environments: a survey on problems, methods and measures

    Soft Comput.

    (2011)
  • Y. Jin et al.

    Evolutionary optimization in uncertain environments – a survey

    IEEE Trans. Evol. Comput.

    (2005)
  • J. Branke, Evolutionary optimization in dynamic...
  • J. Branke et al.

    Designing evolutionary algorithms for dynamic optimization problems

    Advances in Evolutionary Computing

    (2003)
  • K. Trojanowski et al.

    Searching for optima in non-stationary environments

    Evolutionary Computation, 1999. CEC 99. Proceedings of the 1999 Congress on

    (1999)
  • J. Duhain et al.

    Towards a more complete classification system for dynamically changing environments

    Proceedings of the IEEE Congress on Evolutionary Computation

    (2012)
  • M.C. Du Plessis

    Adaptive multi-population differential evolution for dynamic environments

    (2012)
  • R.C. Eberhart et al.

    Tracking and optimizing dynamic systems with particle swarms

    Proceedings of the IEEE Congress on Evolutionary Computation

    (2001)
  • X. Hu et al.

    Tracking dynamic systems with PSO: where’s the cheese

    Proceedings of the Workshop on Particle Swarm Optimization

    (2001)
  • P.J. Angeline

    Tracking extrema in dynamic environments

    Evolutionary Programming VI

    (1997)
  • J. Branke

    Memory enhanced evolutionary algorithms for changing optimization problems

    Proceedings of the IEEE Congress on Evolutionary Computation (CEC 1999), Washington, DC, USA

    (1999)
  • T. Bäck

    On the behavior of evolutionary algorithms in dynamic environments

    1998 IEEE International Conference on Evolutionary Computation Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98TH8360)

    (1998)
  • R.W. Morrison

    Performance measurement in dynamic environments

    GECCO Workshop on Evolutionary Algorithms for Dynamic Optimization Problems

    (2003)
  • K. De Jong

    An analysis of the behavior of a class of genetic adaptive systems

    (1975)
  • C. Li et al.

    Benchmark Generator for CEC 2009 Competition on Dynamic Optimization

    Technical Report

    (2008)
  • W. Feng et al.

    Benchmarks for testing evolutionary algorithms

    Asia-Pacific Conference on Control and Measurement

    (1998)
  • K. Weicker

    Performance measures for dynamic environments

  • T.T. Nguyen

    Continuous dynamic optimisation using evolutionary algorithms

    (2011)
  • K. Beyer et al.

    When is “nearest neighbor” meaningful?

    International Conference on Database Theory

    (1999)
  • C.C. Aggarwal et al.

    On the surprising behavior of distance metrics in high dimensional space

    International Conference on Database Theory

    (2001)
  • C. Perwass et al.

    Algebra

    Geometric Algebra with Applications in Engineering

    (2009)
  • D.J. Sheskin

    Handbook of Parametric and Nonparametric Statistical Procedures

    (2003)
  • J. Demšar

    Statistical comparisons of classifiers over multiple data sets

    J. Mach. Learn. Res.

    (2006)
  • S. Garcia et al.

    An extension on “statistical comparisons of classifiers over multiple data sets” for all pairwise comparisons

    J. Mach. Learn. Res.

    (2008)
  • S. García et al.

    A study on the use of non-parametric tests for analyzing the evolutionary algorithms’ behaviour: a case study on the CEC’2005 special session on real parameter optimization

    J. Heuristics

    (2009)
  • Cited by (6)

    View full text