Skip to main content

Divergence Projections for Variable Selection in Multi–layer Perceptron Networks

  • Conference paper
Neural Nets (WIRN 2003)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2859))

Included in the following conference series:

  • 387 Accesses

Abstract

In this paper an information geometric–based variable selection method for MLP networks is shown. It is based on divergence projections of the Riemannian manifold defined by a MLP network on submanifolds defined by MLP networks with reduced input dimension. We show how we can take advantage of the layered structure of the MLP to simplify the projection operation, which cannot be accurately done by using only the Fisher information metric. Furthermore, we show that our selection algorithm is more robust and gives better results than other well known selection algorithms like Optimal Brain Surgeon. Some examples are shown to validate the proposed approach.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Amari, S.: Information geometry of EM and em algorithms for neural networks. Neural Networks 8, 1379–1408 (1995)

    Article  Google Scholar 

  2. Amari, S.: Methods of Information Geometry. Translations of Mathematical Monographs 191. Oxford University Press, Oxford (2000)

    MATH  Google Scholar 

  3. Dupuis, J.A., Robert, C.P.: Bayesian Variable Selection in Qualitative Models by Kullback–Leibler projections. J. Statistical Planning and Inference 111, 77–94 (2003)

    Article  MATH  MathSciNet  Google Scholar 

  4. Fukunaga, K.: Introduction to Statistical Pattern Recognition, 2nd edn. Academic Press, London (1990)

    MATH  Google Scholar 

  5. Hassibi, B., Stork, D.G.: Second order derivatives for network pruning: Optimal Brain Surgeon. In: Hanson, S.J., Cowanadn, J.D., Giles, C.L. (eds.) Advances in Neural Information Processing Systems 5: Proceedings of the 1992 Conference, pp. 164–171. Morgan Kaufmann, San Mateo (1993)

    Google Scholar 

  6. van de Laar, P., Heskes, T., Gielen, S.: Partial retraining: A new approach to input relevance determination. International Journal of Neural Systems A 9, 75–85 (1999)

    Article  MATH  Google Scholar 

  7. LeCun, Y., Denker, J.S., Solla, S.A.: Optimal Brain Damage. In: Touretzky, D.S. (ed.) Advances in Neural Information Processing Systems 2: Proceedings of the 1989 Conference, pp. 598–605. Morgan Kaufmann, San Mateo (1990)

    Google Scholar 

  8. Leray, P., Gallinari, P.: Feature Selection with Neural Networks. Behaviormetrika (special issue on Analysis of Knowledge Representation in Neural Network Models) 26(1), 145–166 (1999)

    Google Scholar 

  9. MacKay, D.J.C.: A practical Bayesian framework for backpropagation networks. Neural Computation 4(3), 448–472 (1992)

    Article  Google Scholar 

  10. McCullagh, P., Nelder, J.: Generalized Linear Models. Chapman and Hall, London (1989)

    MATH  Google Scholar 

  11. Neal, R.M.: Bayesian Learning for Neural Networks. Springer, New York (1996)

    MATH  Google Scholar 

  12. Stahlberger, A., Riedmiller, M.: Fast network pruning and feature extraction using the unit–OBS algorithm. In: Mozer, M., Jordan, M., Petsche, T. (eds.) Advances in Neural Information Processing Systems 9, pp. 655–661. MIT Press, Cambridge (1997)

    Google Scholar 

  13. Robert, C.P., Casella, G.: Markov Chain Monte Carlo Methods. Springer, New York (1999)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Eleuteri, A., Tagliaferri, R., Milano, L. (2003). Divergence Projections for Variable Selection in Multi–layer Perceptron Networks. In: Apolloni, B., Marinaro, M., Tagliaferri, R. (eds) Neural Nets. WIRN 2003. Lecture Notes in Computer Science, vol 2859. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-45216-4_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-45216-4_32

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-20227-1

  • Online ISBN: 978-3-540-45216-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics