Elsevier

Information Sciences

Volume 165, Issues 3–4, 19 October 2004, Pages 207-220
Information Sciences

Evolving RBF neural networks for time-series forecasting with EvRBF

https://doi.org/10.1016/j.ins.2003.09.025Get rights and content

Abstract

This paper is focused on determining the parameters of radial basis function neural networks (number of neurons, and their respective centers and radii) automatically. While this task is often done by hand, or based in hillclimbing methods which are highly dependent on initial values, in this work, evolutionary algorithms are used to automatically build a radial basis function neural networks (RBF NN) that solves a specified problem, in this case related to currency exchange rates forecasting. The evolutionary algorithm EvRBF has been implemented using the evolutionary computation framework evolving object, which allows direct evolution of problem solutions. Thus no internal representation is needed, and specific solution domain knowledge can be used to construct specific evolutionary operators, as well as cost or fitness functions. Results obtained are compared with existent bibliography, showing an improvement over the published methods.

Introduction

A radial basis function (RBF), φ, can be characterized by a point of the input space, c, and a radius or width, r, such that the RBF reaches its optimum value (maximun/minimun) when applied to c, and decreases/increases to its opposite optimum value when applied to points far from c. The radius, r, controls how distance affects that increment or decrement. For this reason, mathematicians have used groups of RBF to successfully interpolate data. Typical examples of RBF are the Gaussian function (see Fig. 1a) and the multiquadric function (see Fig. 1b), although there are many others [1], [2].

RBF are used by radial basis function neural networks (RBF NNs), which were introduced by Broomhead and Lowe in [3], being their main applications function approximation and time-series forecasting, as well as classification or clustering tasks. Traditionally, a RBF NN is thought as a two-layer, feed-forward network in which hidden neuron activation functions are RBF (see Fig. 2). Very often, the function used is the Gaussian one.

In RBF NNs each hidden neuron computes the distance from its input to the neuron's central point, c, and applies the RBF to that distance, as shows Eq. (1).hi(x)=φ∥x−ci2/ri2where hi(x) is the output yielded by hidden neuron number i when input x is applied; φ is the RBF, ci is the center of the ith hidden neuron, and ri is its radius.

In the next step, the neurons of the output layer perform a weighted sum using the outputs of the hidden layer and the weights of the links that connect both output and hidden layer neurons (Eq. (2)).oj(x)=∑i=0n−1wijhi(x)+w0jwhere oj(x) is the value yielded by output neuron number j when input x is applied; wij is the weight of the links that connects hidden neuron number i and output neuron number j, w0j is a bias for the output neuron, and finally, n is the number of hidden neurons.

It has been proved [4] that when enough units are provided, a RBF NN can approximate any multivariate continuous function as much as desired. This is possible because once the centers and the radii have been fixed, the weights of the links between the hidden and the outputs layers can be calculated analytically using singular value decomposition [5] or any algorithm suitable to solve lineal algebraic equations, making unnecessary the use of training algorithms, as those used in other kinds of neural networks such as multilayer perceptrons. Thus, the main problem in RBF NNs design concerns establishing the number of hidden neurons to use and their centers and radii.

The need of automatic mechanisms to build RBF NNs is already present in Broomhead and Lowe's work [3], where they showed that one of the parameters that critically affects the performance of RBF NNs is the number of hidden neurons. When this number is not sufficient, the approximation offered by the net is not good enough; in the other hand, nets with many hidden neurons will approximate very well those points used to calculate the connection weights, while having a very poor predictive power; this is the so-called overfitting problem [6]. In consequence, establishing the number of neurons (that is, the number of centers and values related to them) that solves a given problem is one of the most important tasks researchers have faced in this field.

In this paper an evolutionary algorithm is used to find the best components of the RBF NNs that approximate a function representing a time-series. Results are compared with other techniques that also use the evolutionary approach to tune the RBF NN.

The rest of the paper is organized as follows: Section 2 describes some of the methods used to solve the cited problems; Section 3 shows our proposed solution, and describes the evolutionary computation framework we use (EO). Next section (4) shows some functional approximation experiments; and finally, our conclusions and future working lines are exposed in Section 5.

Section snippets

State of the art

There are several papers that address the problem of automatic RBF NN design. The paper by Leonardis and Bischof [7] offers a good overview. Methods found in bibliography can be divided in five classes.

(1) Methods in which the number of radial basis functions must be given a priori; after this, computing the values for the centers and the radii is done choosing points randomly from the training set or performing any kind of clustering method with them [8].

(2) Methods that automatically search

EvRBF

The method introduced in this paper uses an evolutionary algorithm, EvRBF, to build RBF NNs, optimizing their generalization error by finding the number of neurons in the hidden layer, and their centers and radii.

The evolutionary algorithm itself has been programmed using the new evolutionary computation framework, evolving objects (EO), current version is 0.9.2 [22], [23]. It can be found at http://eodev.sourceforge.net and is available under open source license.

This new framework is the

Experiments and results

Previous experiments in function approximation using this new approach can be found in [25].

For this new work, the time-series being forecasted is the one used by Sheta and de Jong in [21]. This time-series is composed of real data representing the exchange rates between British pound and US dollar during the period going from 31 December 1979 to 26 December 1983, available from http://pacific.commerce.ubc.ca/xr/data.html, thanks to the work done by Prof. Werner Antweiler, from the University

Conclusions

Creating the best RBF NN that solves a given problem is a difficult task because many parameters have to be set at the same time: number of hidden neurons, centers and radii for them. Evolutionary algorithms can help in finding the optimal values for those parameters, obtaining RBF NNs with low generalization error.

In order to use specific problem knowledge, the RBF NNs are evolved as such, without the typical use of a representation (binary or floating point vector) and a decoder. This is

Acknowledgements

This work has been supported in part by TIC2002-04036-C05-04 and TIC2003-09481-C04.

References (25)

  • S. Chen et al.

    Orthogonal least squares algorithm for learning radial basis function networks

    IEEE Transactions on Neural Networks

    (1991)
  • S. Chen et al.

    Regularized orthogonal least squares algorithm for constructing radial basis function networks

    International Journal of Control

    (1996)
  • Cited by (0)

    View full text