Elsevier

Neurocomputing

Volume 150, Part A, 20 February 2015, Pages 43-49
Neurocomputing

τ-NEAT: Initial experiments in precise temporal processing through neuroevolution

https://doi.org/10.1016/j.neucom.2014.04.077Get rights and content

Abstract

Neuroevolution of Augmenting Topologies (NEAT) has been a very successful algorithm for evolving Artificial Neural Networks (ANNs) that adapt their structure and processing to the task that is required from them. However, this algorithm is not always reliable when handling time related processes and this may be due to its lack of explicit temporal elements within its structure. Of course, NEAT can handle time dependent phenomena through the use of recurrences within the networks it builds, but it is well known that simple recurrences do not easily allow for precise temporal processing due to the history effect they induce on the networks. Many authors have argued for the introduction of other mechanisms, which are also present in natural systems, such as variable or trainable propagation delays in the synapses of the networks that must deal with precise temporal processing. In this paper, we carry out an initial study of a new implementation of NEAT called τ-NEAT that includes the possibility of introducing variable delays in the synapses of the networks NEAT constructs. These delays can affect both, regular direct synapses or recurrent connections. To evaluate the performance of this implementation several tests are carried out over different types of temporal functions and the results of the traditional version of NEAT and τ-NEAT are compared.

Introduction

The NEAT (neuroevolution of augmenting topologies) algorithm [1], [2] is a widely tested approach for evolving the weights and structure of an artificial neural network (ANN). Its operation is based on the use of history markers in genes to promote crossover between similar topologies. Thus, species or niches in the population are preserved by avoiding reproduction between historically different individuals. Moreover, NEAT starts with simple feed-forward ANNs that contain only input and output neurons and it incrementally increases their complexity through structural mutation operators, the add connection mutation and the add node mutation [3]. This way, a designer does not need to predetermine the architecture and number of nodes of the ANN needed for a given task or function, and this is very useful when applying ANNs to problems and domains where it is not easy to predetermine the difficulty of the task. In this type of situations, when the network is too small, the function is learnt without too much detail and large errors may arise. On the other hand, when too many nodes are used, if one is not very careful, overtraining may easily occur leading to a lack of generalization capabilities of the networks in the task. As a consequence, having an algorithm like NEAT that automatically and simultaneously grows and trains a network until its function is achieved with the required goodness is of great importance. In fact, this algorithm has been successfully applied to very different domains going from data classification [4], [5] to evolutionary robotic design [6], but its main application field has been that of learning in dynamic domains, like video games [7], [8], [9] or vehicle crash simulation [10].

Time dependent processing is quite important in many applications and NEAT is able to manage time dependent phenomena through recurrent or feedback connections between neurons that can be inserted using the add connection mutation operator. Consequently, NEAT intrinsically supports the generation and training of classical Recurrent Neural Networks, which are quite adept at working with dynamic processes that depend on sequences of events.

However, classical recurrent neural networks (RNN) present several drawbacks when dealing with problems that require precise timing [11], especially when modeling the underlying structure of complex time series, and different approaches have been developed to address them [11], [12]. One of the most popular consists in mimicking nature and modeling the length of the synapses through the introduction of synaptic time delays both in the direct and in the recurrent connections, leading to what have been called time delay recurrent neural networks (TDRNN) [13], [14], [15].

The use of delays when modeling dynamic processes is supported by a series of theoretical results in the signal processing literature such as the embedding theorem [16], [17]. This theorem states that given an unknown dynamic system:x(n+1)=F(x(n))where x(n) is the d-dimensional state vector of the system and F some function; the information of which we obtain by sampling an observable in time, that is through a temporal series given byY(n)=h(x(n))+r(n)being h() a scalar function and r() a noise term, the geometric structure of the multivariate dynamics of the system in the absence of noise can be extended from observable Y(n) to a D-dimensional space asYR(n)=[Y(n),Y(nτ),,Y(n(D1)τ)]Twhere τ is called the embedding delay.

What is important here is that this theorem states that having an observable Y(n) that corresponds to one component of an unknown dynamical system, the dynamic reconstruction of the signal (in other words a precise model) is possible using vector YR(n) from a dimension D onwards. In fact, it can be said that this dimension must verify:D2d+1which is a sufficient, but not necessary, condition. The minimum dimension D that permits this reconstruction is called the embedding dimension.

Note that what this means is that when we have a dynamic system characterized by a measured signal, it is only necessary to embed this signal in a higher dimensional space of dimension D by taking D samples of the signal spaced by τ in order to make it predictable or, in other words, its unambiguous modeling feasible.

Thus, according to the embedding theorem the only problem now becomes how to obtain D and τ [18]. Translating it to our problem, the challenge is to obtain these values autonomously for a signal or process that is being modeled. Basically, how to obtain the number of points that must be considered and the temporal spacing between them when they are regularly spaced.

Going one step further, one could hypothesize that, in many cases, lower dimensional embedding spaces could be used if the samples were not evenly spaced in time and, consequently, if one considered an uneven distribution of delays, a lower number of points would be necessary to disambiguate many dynamic processes. In fact several authors have already hinted towards this conclusion [19].

Thus, to produce a good intrinsic model of a signal, it is necessary to be able to determine what points of the signal must be processed together. This is done in time delay based neural networks, and, in particular, in TDRNNs by using the synapses as delayers of signals, in other words, the signal transferred from one neuron to another suffers a delay that is proportional to a value that characterizes the synapse connecting the neurons. These delays are a sort of representation of the different lengths these connections could present, which would have a bearing on the time signals would take to traverse them. Therefore, when a neuron is processing its inputs, it is really processing points of the signal that correspond to different times. Several authors have reported quite interesting results using this approach in fields such as robotics [20], [21] or dynamic control [22], [23].

However, the TDRNN related algorithms that have been developed do not provide for these networks to be grown and adapt their topology and weights to the problems they are faced with. That is, the designer must usually decide on the number of neurons and the architecture of the TDRNN and a training algorithm is used to provide values for the synaptic weights and delays. This takes us back the problem of how to determine the right size and connectivity of the network, and an obvious solution would be to adapt NEAT so that it can automatically and incrementally generate TDRNN type networks.

Thus the question that is posed here is whether adding the capability of introducing synaptic delays to NEAT, which leads to an algorithm we have called τ-NEAT, can improve the response of the ANNs it produces. That is, we aim to analyze whether such a higher degree of temporal processing is beneficial for NEAT when applied to tasks that involve complex precise temporal patterns, in particular, to time series prediction tasks or precise signal modeling tasks. In other words, does the introduction of time delays allow NEAT to produce better signal modelers?

The paper is organized as follows. Section 2 deals with the formal description of the τ-NEAT algorithm. Section 3 contains the comparison experiments that have been performed using chaotic temporal series and that show how τ–NEAT outperforms NEAT in these complex cases. Finally, Section 4 is devoted to the presentation of the main conclusions of this study.

Section snippets

The τ-NEAT neuroevolutionary algorithm

To allow for the introduction of time delays, the original NEAT algorithm was extended, thus creating the τ-NEAT algorithm. τ-NEAT is basically a neuroevolutionary algorithm for growing neural networks that may include recurrent connections and synaptic delays. Fig. 1 displays the structure of a general or prototypic neural network that τ-NEAT may obtain, where it can be observed that it now includes a synaptic delay τij, in addition to the synaptic weight wij corresponding to the synapse

Experimental analysis of τ-NEAT

To analyze the prediction performance of the τ-NEAT approach, in this work a set of chaotic time series prediction experiments have been carried out. Three chaotic time series have been selected to compare the performance of the new proposal with the performance of the traditional NEAT approach. These time series are the logistic map, the Henon attractor and the Mackey–Glass time series. The first one was chosen due to the fact that its state space can be plotted in two or three-dimensional

Conclusions

The precision necessary to implement an ANN that is able to model a chaotic time series is in general hard to achieve when using ANNs whose only time processing capability is through recurrences. In fact, when one wants to use neuroevolutionary algorithms, such as NEAT, which produce both the topology and the weights of a neural network, thus supposedly obtaining the best network for a given task, one finds that their time processing toolbox is limited to recurrences and thus they are usually

Pilar Caamaño is a postdoctoral researcher at the University of A Coruña, Spain. She received the B.S. and M.S. degree in Computer Science from the University of A Coruña in 2006 and a Ph.D. in Computer Science in 2011. She is a member of the Integrated Group for Engineering Research at the University of A Coruña. Current research activities are related to evolutionary algorithms and fitness landscapes formal analysis.

References (25)

  • G. Wang et al.

    The application of improved NeuroEvolution of Augmenting Topologies neural network in Marcellus Shale lithofacies prediction

    Comput. Geosci.

    (2013)
  • Sung-Suk Kim

    Time-delay recurrent neural network for temporal correlations and prediction

    Neurocomputing

    (1998)
  • K.O. Stanley et al.

    Evolving neural networks through augmenting topologies

    Evol. Comput.

    (2002)
  • K.O. Stanley, R. Miikkulainen, Efficient evolution of neural networks topologies, in: Proceedings of the 2002 Congress...
  • K.O. Stanley et al.

    Competitive coevolution through evolutionary complexification

    J. Artif. Intell. Res.

    (2004)
  • L. Chen, D. Alahakoon, NeuroEvolution of augmenting topologies with learning for data classification, in:Second...
  • P. Krčah , Towards efficient evolution of morphology and control, GECCO’08, in: Proceedings of the 10th Annual...
  • K.O. Stanley et al.

    Real-time neuroevolution in the NERO video game

    IEEE Trans. Evol. Comput.

    (2005)
  • W.L. Raffe et al.

    Neuroevolution of content layout in the PCG: angry bots video game, 2013

    IEEE Congress on Evolutionary Computation, CEC 2013

    (2013)
  • L. Cardamone, D. Loiacono, P.L. Lanzi, Evolving competitive car controllers for racing games with neuroevolution, in:...
  • N. Kohl, K. Stanley, R. Miikkulainen, M. Samples, R. Sherony, Evolving a real-world vehicle warning system, in:GECCO...
  • F.A. Gers et al.

    Learning precise timing with lstm recurrent networks

    J. Mach. Learn. Res.

    (2003)
  • Cited by (6)

    Pilar Caamaño is a postdoctoral researcher at the University of A Coruña, Spain. She received the B.S. and M.S. degree in Computer Science from the University of A Coruña in 2006 and a Ph.D. in Computer Science in 2011. She is a member of the Integrated Group for Engineering Research at the University of A Coruña. Current research activities are related to evolutionary algorithms and fitness landscapes formal analysis.

    Francisco Bellas is a Profesor Titular at the University of A Coruña, Spain. He received the B.S. and M.S. degree in Physics from the University of Santiago de Compostela, Spain, in 2001, and a Ph.D. in Computer Science from the University of A Coruña in 2003. He is a member of the Integrated Group for Engineering Research at the University of A Coruña. His main research activities are focused to evolutionary robotics, collective intelligence and neuroevolutionary algorithms.

    Richard J. Duro received a M.S. degree in Physics from the University of Santiago de Compostela, Spain, in 1989, and a Ph.D. in Physics from the same University in 1992. He is currently a Catedrático de Universidad in the Department of Computer Science and head of the Integrated Group for Engineering Research at the University of A Coruña. His research interests include higher order neural network structures, signal processing and autonomous and evolutionary robotics.

    View full text