On-line RBFNN based identification of rapidly time-varying nonlinear systems with optimal structure-adaptation

https://doi.org/10.1016/S0378-4754(02)00159-3Get rights and content

Abstract

This paper presents an adaptive RBF network for the on-line identification and tracking of rapidly-changing time-varying nonlinear systems. The proposed algorithm is capable of maintaining the accuracy of learned patterns even when a large number of aged patterns are replaced by new ones through the adaptation process. Moreover, the algorithm exhibits a strong learning capacity with instant embodiment of new data which makes it suitable for tracking of fast-changing systems. However, the accuracy and speed in the adaptation is balanced by the computational cost which increases with the square of the number of the radial basis functions, resulting in a computational expensive, but still practically feasible, algorithm. The simulation results show the effectiveness (in terms of degradation of learned patterns and learning capacity) of this architecture for adaptive modeling.

Introduction

Following their broad learning and modeling capabilities, artificial neural networks are established as primary tools for optimization, pattern recognition and nonlinear system identification [1], [4], [6], [12] and control [5], [13], [14], [15]. Radial basis function (RBF) networks are three-layered neural networks composed of nodes whose output is proportional to the distance from the node center. Broomhead and Lowe [1] first studied the RBF networks as nonlinear function estimators and indicated their interpolation capabilities. Hartman et al. [2] and later Park and Sandberg [9], [10] proved that RBFs are capable of approximating any function with arbitrary accuracy. Since then, substantial efforts have been made to establish the learning efficiency and convergence rates of RBFs [4], [7], [16].

RBF networks have emerged as an alternative approach to multi-layer perceptrons (MLP) for pattern recognition and nonlinear systems modeling. This is due to the fast two-stage learning algorithms developed for RBFs in contrast to the slow convergence of the back-propagation algorithm which is widely used for the training of MLPs [6]. In the first phase of these popular RBF training methods, the centers and widths of the nodes of the network are selected. This could be done either randomly according to the training data or by employing an unsupervised procedure. The incentive behind this is that by placing the node centers following the training data density, near-optimum coverage of the input space can be expected. Various clustering algorithms for the choice of centers and widths have been applied successfully to many problem domains. In the second phase the output weights can be calculated using simple supervised least mean squares procedures since the network output is linear with respect to the weights. The decoupling could cause in some cases some loss of information although these algorithms are capable to compensate this loss by usually having significant smaller computational cost. Two-stage procedures often lead to better solutions given limited training data and computational time, although they can scarcely reach an optimal solution even if they go through intensive training.

Supervised training involving simultaneous adaptation of all the RBF network parameters attracted less research interest and not until recently have researchers started to study more systematically this domain [3]. The localized structure of RBF networks inevitably leads to inferior generalization performance compared to sigmoidal MLP neural networks. This constraint has up to now being tackled with the above mentioned methods of unsupervised selection of RBF centers and widths which are more capable of capturing the modeled function’s input space structure. This implies that in order for a supervised trained RBFNN to achieve similar to MLPNN interpolation capabilities, it needs significantly more computational resources.

The problem formulation is briefly discussed in Section 2. The proposed methodology is presented in Section 3. Section 4 discusses the implementation details and in Section 5 the mathematical formulation of the model is given. Some simulation results which show the model effectiveness are given in Section 6. Finally, Section 7 contains some concluding remarks.

Section snippets

Problem formulation

Adaptation to uncertain or time-varying dynamics has always been a challenge in the area of neural networks. The majority of training algorithms used for system identification requires the whole input–output data set to operate, rendering them totally unsuitable for modeling of nonstationary systems. Of course, on-line or incremental training of neural networks has been tackled by numerous researchers [8], [11], [17] but these approaches are also not extendable to time-varying dynamics. This is

Proposed model concept

The adaptive model described in this paper is used for identification and tracking of nonlinear time-varying systems. The data needed to train this model are the pairs of the system input and corresponding output without any other information about the system necessary. The key feature of our method compared to others is the direct approach used for formulating the training target. The training target precisely reflects the optimal input–output mapping after a new input–output pair (which is

Implementation

Each input–output pair is processed individually as soon as it becomes available and after that it is discarded, thus eliminating the need for memory of past data. Using the new data pair, a target model is constructed. The target model is the actual model changed such that to include the new information. Finally, the actual model is trained by a gradient descent procedure to match the target model.

In this paper the target and actual models are implemented as RBF networks with Gaussian basis

Mathematical formulation

For simplicity of the representation, the mathematical formulation is restricted to the single-output case (Fig. 3). The extension to multiple-output systems is straightforward.

Consider a nonlinear time-varying system with M inputs defined by the vector:x=x1x2xMTand one output ys. The system is described by the following equation:ys=F(x,t)The goal of the neural model described in this paper is to track the above system so that at each time the model mapping is as close as possible to the

Simulation results

Here, three application examples of the proposed model are provided which illustrate its capabilities and effectiveness. The RBF network used in all of the examples has 20 nodes in the hidden layer.

Concluding remarks

This paper has presented a new method for the modeling of time-varying nonlinear systems. The approach adopted is based on a combination of actual and target models. The proposed scheme has the following advantages which are vital for the modeling of time-varying systems: (i) persistent memory of learned structures; (ii) fast learning of new data without degradation of learned mapping; and (iii) immunity to model parameter shifting or overtraining phenomena. Moreover, the application of this

References (17)

  • L. Xu et al.

    On radial basis function nets and kernel regression: approximation ability, convergence rate and receptive field size

    Neural Netw.

    (1994)
  • D.S. Broomhead et al.

    Multivariable function interpolation and adaptive networks

    Complex Syst.

    (1988)
  • E.J. Hartman et al.

    Layered neural networks with Gaussian hidden units as universal approximators

    Neural Comput.

    (1990)
  • N.B. Karayiannis

    Reformulated radial basis neural networks trained by gradient descent

    IEEE Trans. Neural Netw.

    (1999)
  • A. Krzyzak et al.

    Radial basis function networks and complexity regularization in function learning

    IEEE Trans. Neural Netw.

    (1998)
  • C.T. Lin, Neural Fuzzy Control Systems with Structure and Parameter Learning, World Scientific, Singapore,...
  • J. Moody et al.

    Fast learning in locally-tuned processing units

    Neural Comput.

    (1989)
  • P. Niyogi et al.

    On the relationship between generalization error, hypothesis complexity, and sample complexity for radial basis functions

    Neural Comput.

    (1996)
There are more references available in the full text version of this article.

Cited by (20)

  • Integration of support vector regression and annealing dynamical learning algorithm for MIMO system identification

    2011, Expert Systems with Applications
    Citation Excerpt :

    In the neural network, RBFNs have received considerable applications, such as function approximation, prediction, recognition, etc. (Chuang, Jeng, & Lin, 2004; Sing, Basu, Nasipuri, & Kundu, 2007; Xu, Xie, Tang, & Ho, 2003; Yu, Gomm, & Williams, 2000). Since RBFNs have only one hidden layer and have fast convergence speed, they are widely used for nonlinear system identification recently (Apostolikas & Tzafestas, 2003; Chen, Hong, Luk, & Harris, 2009; Falcao, Langlois, & Wichert, 2006; Fu et al., 2009; Li & Zhao, 2006). Besides, the RBFNs are often referred to as model-free estimators since they can be used to approximate the desired outputs without requiring a mathematical description of how the outputs functionally depend on the inputs (Kosko, 1992).

  • Radial basis function networks with hybrid learning for system identification with outliers

    2011, Applied Soft Computing Journal
    Citation Excerpt :

    Radial basis function networks (RBFNs) have been widely used for various applications, such as function approximate, prediction, recognition, classification [5,21,25,29]. Since the RBFNs have only one hidden layer and have fast convergence speed [23], they are widely used for the nonlinear system identification recently [1,2,16]. Besides, the RBFNs are often referred to as model-free estimators since they can be used to approximate the desired outputs without requiring a mathematical description of how the outputs functionally depend on the inputs [14].

View all citing articles on Scopus
View full text