Elsevier

Neurocomputing

Volume 71, Issues 7–9, March 2008, Pages 1203-1209
Neurocomputing

A multi-objective approach to RBF network learning

https://doi.org/10.1016/j.neucom.2007.11.021Get rights and content

Abstract

The problem of inductive supervised learning is discussed in this paper within the context of multi-objective (MOBJ) optimization. The smoothness-based apparent (effective) complexity measure for RBF networks is considered. For the specific case of RBF network, bounds on the complexity measure are formally described. As the synthetic and real-world data experiments show, the proposed MOBJ learning method is capable of efficient generalization control along with network size reduction.

Introduction

The supervised learning paradigm has been widely discussed in recent years. From the statistical learning theory, it is known that conditions for a good generalization can be reached at the minimum of the true risk. However, this turns out to be difficult to achieve since only empirical risk can be minimized directly. Non-synthetic data are most likely to be finite and disturbed by noise, often leading direct empirical risk minimization methods to a poor generalization and over-fitting. The key to a good generalization lies in the control of the capacity (complexity) of a learning machine that is formulated by the well-known structural risk minimization (SRM) principle [21]. The search for a proper complexity can be also viewed as a problem of balance between bias and variance [8]. Pruning methods [15], regularization networks [9] and support vector machines [20] (which are a certain form of regularization) are commonly used to achieve that goal. While pruning algorithms control the complexity by manipulating a network structure (e.g. number of nodes), regularization methods aim at controlling the network output response in spaces of smoothness [19], manipulating the so-called apparent complexity. Being a more general complexity representation, the apparent complexity as measure of smoothness may restrict the learning capabilities of the structure. Another example of apparent complexity is the known relationship between learning capacity of multi-layer perceptrons (MLP) and the size of its weights [3].

The approaches mentioned above jointly minimize network complexity and empirical error in the form of a single loss-function that usually consists of an error cost function and a regularizer. Although this may result in good generalization models, they are highly dependent on user defined training parameters. In addition, it is generally known that error and complexity are conflicting objectives and, similarly to bias and variance, they demand trading-off instead of the joint minimization. This viewpoint led to the development of multi-objective machine learning (MOML) methods [10], which treat the empirical risk and learning capacity of a hypothesis class as two separate objectives. The evolutionary algorithms [11] are commonly used for solving MOML problems, however [4], [6], [16] show the efficiency of multi-objective (MOBJ) algorithms based on nonlinear programming.

Consider a neural network of a certain structure with its input–output response determined by the parameter vector ωΩ. Introducing error and complexity objective functions φe(ω) and φc(ω), respectively, with the first one representing the empirical risk and the second one representing the learning capacity, we may now formulate the vector optimization problem for MOML asminωΩ[φe(ω),φc(ω)].Since the objectives are conflicting in the region of interest, the solution of (1) results in a set of Pareto-optimal solutions Ω*={ω*Ω:φe(ω*)φe(ω)and φc(ω*)φc(ω),ωΩ},which form the so-called Pareto front of the feasible region. In other words, the Pareto front Ω* contains the solutions which represent the best compromise between the two objectives. It means that for each solution ωΩ* there always exists at least one solution ω*Ω* having lower complexity and error. As common, the squared error criterion and the norm of network weights w are taken as error and complexity measures for MLP networks [16]. To the authors’ knowledge, a general definition for assessing apparent complexity of other network types is not known, that is an obstacle for MOBJ learning. In this paper we present the smoothness-based apparent complexity measure and determine its bounds for radial basis function (RBF) neural networks. In particular, we show further that the apparent complexity can be limited by the ratio of the 1-norm of the weights vector to the width of the RBFs. It is also demonstrated that such form of the complexity control leads to sparse weights, eventually making the proposed MOBJ learning method to be capable of reducing the network size.

Section snippets

Apparent complexity measure for RBF networks

Consider the smooth input–output mapping function f:XR with the compact support in XRn. In general, one can express the smoothness of f in Sobolev space [1] Wk,p(X), endowed with the normfk,p=0|r|kDrfpp1/por its equivalent normfk,p=fp+|r|=kDrfp.Here Drf(x)=|r|x1r1x2r2xnrn,D0f=f(x)is the rth order generalized partial derivative operator, where r=(r1,r2,,rn) and |r|=i=1nri, and the Lp-norm is defined as fp=X|f(x)|px, p1.

In particular case for p=2, the spaces Wk,2 are

MOBJ learning

Usually, the vector optimization problem (1) is difficult to solve analytically; however, in practice it is enough to approximate the Pareto front Ω* with a finite number of solutions. According to MOBJ learning concept, when a certain approximation is obtained, the resulted “best” solution must be selected from Ω* on a decision making step, with respect to a posteriori model selection criterion (e.g. minimum validation error or maximum entropy).

A variety of methods for the obtaining of

Experiments and discussion

In this section, we present further several experiments on both the synthetic and real-world data. All experiments follow the same MOBJ learning scheme described in Section 3. The classical ellipsoid method was applied for solving the optimization problem (10).

Conclusions

In this work, we derive the bounds on smoothness-based apparent complexity for RBF networks exploring a formulation of the smoothness in Sobolev spaces. It is confirmed experimentally that the proposed bounds participating in the MOBJ scheme provide the efficient solutions of learning problems.

The proposed MOBJ method controls the generalization in loss-smoothness spaces similarly to regularization learning, but in addition, it provides a complete solution search, regardless of the convexity of

Acknowledgements

The authors are grateful to CNPq, Brazil for financial support.

Illya Kokshenev received the M.S. degree in Computer Science from the Kharkov National University of Radioelectronics (Ukraine) in 2004. He is currently a Ph.D. student at the Department of Electronic Engineering, Federal University of Minas Gerais (Brazil). His research interests cover machine learning, neural networks, neuro-fuzzy systems and, presently, multi-objective machine learning.

References (21)

  • M.A. Costa et al.

    Training neural networks with a multi-objective sliding mode control algorithm

    Neurocomputing

    (2003)
  • R.A. Adams et al.

    Sobolev Spaces

    (2003)
  • A. Asuncion, D. Newman, UCI machine learning repository. URL 〈http://www.ics.uci.edu/∼mlearn/MLRepository.html〉,...
  • P.L. Bartlett

    For valid generalization the size of the weights is more important than the size of the network

  • A.P. Braga et al.

    Multi-objective algorithms for neural-networks learning

  • M.A. Costa, A.P. Braga, Optimization of neural networks with multi-objective lasso algorithm, in: International Joint...
  • B. Efron, T. Hastie, I. Johnstone, R. Tibshirani, Least angle regression, Technical Report, Department of Statistics,...
  • S. Geman et al.

    Neural network and the bias/variance dilemma

    Neural Comput.

    (1992)
  • F. Girosi et al.

    Regularization theory and neural networks architectures

    Neural Comput.

    (1995)
There are more references available in the full text version of this article.

Cited by (28)

  • A hybrid multiobjective RBF-PSO method for mitigating DoS attacks in Named Data Networking

    2015, Neurocomputing
    Citation Excerpt :

    Some of the existing algorithms to achieve that are given in [32,36,41–43]. Almost all algorithms compute the optimal estimation of the basis function centers by means of error minimization, i.e., accuracy based on Mean-Square Error (MSE) [36,43–45]. However, MSE is not suitable for determining the optimal position of basis function centers.

  • Reinforcement radial basis function neural networks with an adaptive annealing learning algorithm

    2013, Applied Mathematics and Computation
    Citation Excerpt :

    In addition, RBFNNs can be used to approximate desired outputs without requiring a mathematical description of their functional dependence on the inputs and are therefore often referred to as model-free estimators [12]. Recently, several researchers proposed efficient learning algorithms for RBFNNs [13–18]. Besides, a number of sequential learning algorithms for resource allocation networks have been developed to improve training error convergence and computational efficiency in the extant literature [19–26].

  • Multi-objective hybrid evolutionary algorithms for radial basis function neural network design

    2012, Knowledge-Based Systems
    Citation Excerpt :

    In addition, the researchers in [30] used multi-objective optimization to find the best of RBF, number of hidden units, centers and widths of RBFNs. The previous works [21–30] have been proposed for unsupervised learning or supervised learning based RBFN separately. However, the current work has developed hybrid learning (unsupervised learning and supervised learning) and structure optimization simultaneously based on proposed methods to obtain simple and accurate RBFNs.

View all citing articles on Scopus

Illya Kokshenev received the M.S. degree in Computer Science from the Kharkov National University of Radioelectronics (Ukraine) in 2004. He is currently a Ph.D. student at the Department of Electronic Engineering, Federal University of Minas Gerais (Brazil). His research interests cover machine learning, neural networks, neuro-fuzzy systems and, presently, multi-objective machine learning.

Antonio Padua Braga was born in Brazil in 1963. He obtained his first degree in Electrical Engineering and his Master's in Computer Science, both at Federal University of Minas Gerais, Brazil. The main subject of his Ph.D. which was received from University of London, England, was Storage Capacity of Boolean Neural Systems. After finishing his Ph.D. he returned to his position as a professor at Department of Electronics Engineering at Federal University of Minas Gerais. He has published several papers in International Journals, Conferences and has written two books in NN. He is the head of the Computational Intelligence Laboratory at Department of Electronics and was co-editor in chief of the International Journal of Computational Intelligence and Applications (IJCIA), published by Imperial College Press. His current research interest areas are the development of learning algorithms for NNs, hybrid neural systems and applications of NNs.

View full text