Skip to main content

Dynamic Construction of Multilayer Neural Networks for Classification

  • Conference paper
Advances in Neural Networks – ISNN 2011 (ISNN 2011)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6675))

Included in the following conference series:

Abstract

There are several drawbacks of multilayer neural networks (MLNNs) including the difficulty of determining the number of hidden nodes and their black box nature. We propose a new dynamic construction mechanism for MLNNs to overcome such inherent drawbacks. The main goal of our work is to train a hidden neuron and assemble it to the network dynamically while making the learning error smaller and smaller. In this paper, a hidden neuron carries out the function of a linear classifier which answers yes(Y) or no(N) to whether the input data belongs to the specific class. We call such a linear classifier a Y/N classifier and call the hidden neuron a Y/N neuron. The number of Y/N neurons are determined self-adaptively according to the given learning error and then successfully avoid the overlearning problem. The dynamically constructed MLNN with Y/N neurons is called a Y/N neural network. We prove that a Y/N neural network can always converge to the required solution and illustrate that Y/N neural networks can be applied to very complex classification problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hornik, K., Stinchcombe, M., White, H.: Multilayer feedforward networks are universal approximators. Neural Networks 2, 359–366 (1989)

    Article  Google Scholar 

  2. Mirchandani, G., Cao, W.: On hidden nodes for neural nets. IEEE Transactions on Circuits and Systems 36, 661–664 (1989)

    Article  MathSciNet  Google Scholar 

  3. Geman, S., Bienenstock, E., Doursat, R.: Neural Networks and the Bias/Variance Dilemma. Neural Computation 4, 1–58 (1992)

    Article  Google Scholar 

  4. Lawrence, S., Giles, C.L., Tsoi, A.C.: What Size Neural Network Gives Optimal Generalization? In: Convergence Properties of Backpropagation. Institute for Advanced Computer Studies, University of Maryland, College Park, MD 20742 (1996)

    Google Scholar 

  5. Bartlett, P.L.: The sample complexity of pattern classification with neural networks: the size of the weights is more important than the size of the network. IEEE Transaction on Information Theory 44, 525–536 (1998)

    Article  MathSciNet  MATH  Google Scholar 

  6. Ash, T.: Dynamic Node Creation in Backpropagation Networks. Connection Science 1, 365–375 (1989)

    Article  Google Scholar 

  7. Elisseeff, A., PaugamMoisy, H.: Size of multilayer networks for exact learning: analytic approach. In: Michael, I., Jordan, M.C.M., Petsche, T. (eds.) Advances in Neural Information Processing Systems, vol. 9, pp. 162–168. The MIT Press, Cambridge (1997)

    Google Scholar 

  8. Fletcher, L., Katkovnik, S.F., Engelbrecht, A.: Optimizing the number of hidden nodes of a feedforward artificial neural network. In: IEEE International Joint Conference on Neural Networks, vol. 2, pp. 1608–1612 (1998)

    Google Scholar 

  9. Sarle, W.S.: Stopped training and other remedies for overfitting. In: 27th Symposium on the Interface, pp. 352–360 (1995)

    Google Scholar 

  10. Xu, L.: Bayesian Ying-Yang System and Theory as A Unified Statistical Learning Approach (III) Models and Algorithms for Dependence Reduction, Data Dimension Reduction, ICA and Supervised Learning. In: Wong, K.M., King, I., Yeung, D.-Y. (eds.) Theoretical Aspects of Neural Computation: A Multidisciplinary Perspective, pp. 43–60. Springer, Heidelberg (1997)

    Google Scholar 

  11. Xu, S., Chen, L.: A novel approach for determining the optimal number of hidden layer neurons for FNN’s and its application in data mining. In: 5th International Conference on Information Technology and Applications, pp. 683–686 (2008)

    Google Scholar 

  12. Vapnik, V.: Support-vector networks. Machine Learning 20, 273–297 (1995)

    MATH  Google Scholar 

  13. Platt, J.: Sequential minimal optimization: A fast algorithm for training support vector machines. Advances in Kernel MethodsSupport Vector Learning 208, 1–21 (1998)

    Google Scholar 

  14. Valiant, L.G.: A theory of the learnable. Communications of the ACM 27, 1134–1142 (1984)

    Article  MATH  Google Scholar 

  15. Schapire, R.E.: The Strength of Weak Learnability. In: 30th Annual Symposium on Foundations of Computer Science, pp. 28–33 (1989)

    Google Scholar 

  16. Schapire, R.E., Freund, Y., Bartlett, P., Lee, W.S.: Boosting the margin: a new explanation for the effectiveness of voting methods. The Annals of Statistics 26, 1651–1686 (1998)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Liu, J., Jia, Y. (2011). Dynamic Construction of Multilayer Neural Networks for Classification. In: Liu, D., Zhang, H., Polycarpou, M., Alippi, C., He, H. (eds) Advances in Neural Networks – ISNN 2011. ISNN 2011. Lecture Notes in Computer Science, vol 6675. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-21105-8_59

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-21105-8_59

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-21104-1

  • Online ISBN: 978-3-642-21105-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics