Skip to main content

Improving RBF networks by the feature selection approach EUBAFES

  • Part III: Learning: Theory and Algorithms
  • Conference paper
  • First Online:
Artificial Neural Networks — ICANN'97 (ICANN 1997)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1327))

Included in the following conference series:

Abstract

The curse of dimensionality is one of the severest problems concerning the application of RBF networks. The number of RBF nodes and therefore the number of training examples needed grows exponentially with the intrinsic dimensionality of the input space. One way to address this problem is the application of feature selection as a data pre processing step.

In this paper we propose a two-step approach for the determination of an optimal feature subset: First, all possible feature-subsets are reduced to those with best discrimination properties by the application of the fast and robust filter technique EUBAFES. Secondly we use a wrapper approach to judge, which of the pre-selected feature subsets leads to RBF networks with least complexity and best classification accuracy. Experiments are undertaken to show the improvement for RBF networks by our feature selection approach.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Berthold MR., Diamond J., Boosting the Performance of RBF Networks with Dynamic Decay Adjustment, Advances in Neural Information Processing Systems, vol.7 (1995)

    Google Scholar 

  2. John G., Kohavi R., Pfleger K., Irrelevant Features and the Subset Selection Problem, Machine Learning: Proceedings of the Eleventh International Conference, (1993) 121–129 William W. Cohne and Haym Hirsh

    Google Scholar 

  3. Kira K., Rendell L.: A practical approach to feature selection. Proceedings of the International Conference on Machine Learning, Aberdeen, (1992) 249–256,Sleeman D., Edwards P. Morgan Kaufmann

    Google Scholar 

  4. Kononenko I., Estimation attributes: Analysis and extensions of RELIEF in Proceedings of the European Conference on Machine Learning, Catana, Italy (1994) 171–182 Springer Verlag”

    Google Scholar 

  5. Moody J. and Darken C.J., Fast learning in networks with locally-tuned processing units Neural Computation, 1, (1989) 281–294

    Google Scholar 

  6. Scherf M., Brauer W., Feature Selection by Means of a Feature Weighting Approach. Technical Report No. FKI-221-97, Forschungsberichte künstliche Intelligenz, Institut für Informatik, Technische Universität München (1997)

    Google Scholar 

  7. Wettschereck D., Aha D.W., Mohori T., A Review and Empirical Evaluation of Feature Weighting Methods for a Class of Lazy Learning Algorithms. Artificial Intelligence Review (to appear) (1997)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Wulfram Gerstner Alain Germond Martin Hasler Jean-Daniel Nicoud

Rights and permissions

Reprints and permissions

Copyright information

© 1997 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Scherf, M., Brauer, W. (1997). Improving RBF networks by the feature selection approach EUBAFES. In: Gerstner, W., Germond, A., Hasler, M., Nicoud, JD. (eds) Artificial Neural Networks — ICANN'97. ICANN 1997. Lecture Notes in Computer Science, vol 1327. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0020186

Download citation

  • DOI: https://doi.org/10.1007/BFb0020186

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-63631-1

  • Online ISBN: 978-3-540-69620-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics