Elsevier

Neurocomputing

Volume 20, Issues 1–3, 31 August 1998, Pages 97-110
Neurocomputing

Normalized Gaussian Radial Basis Function networks

https://doi.org/10.1016/S0925-2312(98)00027-7Get rights and content

Abstract

The performances of normalised RBF (NRBF) nets and standard RBF nets are compared in simple classification and mapping problems. In normalized RBF networks, the traditional roles of weights and activities in the hidden layer are switched. Hidden nodes perform a function similar to a Voronoi tessellation of the input space, and the output weights become the network's output over the partition defined by the hidden nodes. Consequently, NRBF nets lose the localized characteristics of standard RBF nets and exhibit excellent generalization properties, to the extent that hidden nodes need to be recruited only for training data at the boundaries of class domains. Reflecting this, a new learning rule is proposed that greatly reduces the number of hidden nodes needed in classification tasks. As for mapping applications, it is shown that NRBF nets may outperform standard RBFs nets and exhibit more uniform errors. In both applications, the width of basis functions is uncritical, which makes NRBF nets easy to use.

Introduction

Normalized radial basis functions (NRBF) differ from standard radial basis functions (RBF) by a seemingly minor modification of their equation (Section 2). This results in novel computational properties which have attracted little attention in the neural network community. Moody and Darken [11]were first to mention normalised RBF nets without elaborating on their functional significance. However, Servin and Cuevas (1993) noted that normalization gave RBF nets the “same classification properties as nets using sigmoid functions”. Cha and Kassam (1995) proposed that “a normalized Gaussian basis function features either localized behaviour similar to that of a Gaussian or nonlocalized behavior like that of a sigmoid, depending on the location of its centre”. Rao et al. [13]interpreted NRBF nets as mixture of expert models and Jang and Sun [7]saw similarities with fuzzy inference systems. These multiple views reflect the fact that NRBF nodes in the hidden layer behave more like case indicators rather than basis functions proper, as is elaborated in Section 2. This property leads to excellent performances in classification tasks, as shown in Section 3. One of the key features of NRBF nets is their excellent generalization, a property that can be exploited to reduce the number of hidden nodes in classification tasks. This is achieved by using a new learning rule proposed in Section 4that is demonstrated in classification and mapping examples. NRBF nets have also given very good results in another class of application, trajectory learning in robotics 1, 6.

Section snippets

Normalized radial basis function networks

Standard radial basis function (RBF) nets (Fig. 1) comprise a hidden layer of RBF nodes and an output layer with linear nodes 4, 11The function of these nets is given byyi(x)=j=1nwijφ(x−xj),where yi is the activity of the output node i, φ(xxj) is the activity of the hidden node j, with a RBF function centred on the vector xj, x is the actual input vector and wij are the weights from the RBF nodes in the hidden layer to the linear output node. Such a net is a universal function approximator

Standard training procedure

For the results shown in this section, training is done in a standard way [3], (p. 170), by recruiting hidden nodes in the first epoch, then, in subsequent epochs, adjusting the positions of the centres of the nodes and the weights to the output node to minimize the output error. Normalised RBF nets and standard RBF nets are trained with the same procedure:

(i) Recruiting a new hidden node centred on an input vector that was beyond a radius of 0.5σ from the centre of an existing node, or slowly

Modified training procedure

The good interpolation and extrapolation properties shown in the previous section suggests that hidden nodes may need to be recruited only in crucial points, close to boundaries between two classes. To verify this hypothesis, the training procedure in Section 3.1was modified so that no new nodes are recruited if the network indicates the correct class by using existing nodes. The details of the modified procedure are:

(i) Check if the output vector y of the net is correct, i.e. |yydesired|<

Conclusion

These initial results show that normalised RBF nets have very good generalisation properties that are beneficial in classification tasks. This is due to the property of normalised RBF nets to produce a significant output even for input vectors far from the centre of the receptive field of any of the hidden nodes.

Taking advantage of this, a modified learning procedure has been proposed in which hidden nodes are recruited only when neighbouring nodes do not point to the same output value. The

Acknowledgements

Stimulating comments and suggestions by Kaspar Althoefer and anonymous referees are gratefully acknowledged and have helped to improve the final version of this paper.

Guido Bugmann was born in 1953 and has two children. He studied Physics at the University of Geneva in Switzerland. In 1986 he completed his PhD on “Fabrication of photovoltaic solar cells with a-Si : H produced by anodic deposition in a DC plasma”. He then worked at the Swiss Federal Institute of Technology in Lausanne on the development of a measurement system using an ultra-sound beam and neural networks to measure the size of air bubbles in bacterial cultures. In 1989, he joined the

References (14)

  • K. Althoefer, G. Bugmann, Planning and learning goal-directed sequences of robot-arm movements, in: F. Fogelman-Soulié,...
  • H.C. Andersen, A. Lotfi, L. Westphal, Comments on functional equivalence between radial basis function networks and...
  • C.M Bishop

    Neural Networks for Pattern Recognition

    (1995)
  • D.S Broomhead et al.

    Multivariable functional interpolation and adaptive networks

    Complex Systems

    (1988)
  • M Brown et al.

    Neurofuzzy Adaptive Modelling and Control

    (1994)
  • G. Bugmann, K.L. Koay, N. Barlow, M. Phillips, D. Rodney, Stable encoding of robot trajectories using normalised radial...
  • J.-S.R. Jang, C.-T. Sun, Functional equivalence between radial basis function networks and fuzzy inference systems,...
There are more references available in the full text version of this article.

Cited by (132)

View all citing articles on Scopus

  1. Download : Download full-size image
Guido Bugmann was born in 1953 and has two children. He studied Physics at the University of Geneva in Switzerland. In 1986 he completed his PhD on “Fabrication of photovoltaic solar cells with a-Si : H produced by anodic deposition in a DC plasma”. He then worked at the Swiss Federal Institute of Technology in Lausanne on the development of a measurement system using an ultra-sound beam and neural networks to measure the size of air bubbles in bacterial cultures. In 1989, he joined the Fundamental Research Laboratories of NEC in Japan and modelled the function of biological neurons in the visual system. In 1992 he joined Prof. John G. Taylor at King's College London to develop applications of the pRAM neuron model and develop a theory of visual latencies. In 1993 he joined the group of Prof. Mike Denham at the University of Plymouth (UK) where he is developing artificial vision systems for robots and investigates path-planning and spatial memory.

Dr. Bugmann has 3 patents and over 70 publications. He is member of the Swiss Physical Society, The Neuroscience Society and The British Machine Vision Association.

View full text