Skip to main content

Learning Capability: Classical RBF Network vs. SVM with Gaussian Kernel

  • Conference paper
  • First Online:
Developments in Applied Artificial Intelligence (IEA/AIE 2002)

Abstract

The Support Vector Machine (SVM) has recently been introduced as a new learning technique for solving variety of real-world applications based on statistical learning theory. The classical Radial Basis Function (RBF) network has similar structure as SVM with Gaussian kernel. In this paper we have compared the generalization performance of RBF network and SVM in classification problems. We applied Lagrangian differential gradient method for training and pruning RBF network. RBF network shows better generalization performance and computationally faster than SVM with Gaussian kernel, specially for large training data sets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Leone, R. D., Capparuccia, R., and Merelli, E.: A Successive Overrelaxation Back-propagation Algorithm for Neural-Network Traning. IEEE Trans. Neural Networks, Vol. 9, No. 3, pp. 381–388 (1998).

    Article  Google Scholar 

  2. Hsin, H.-C., Li, C.-C., Sun, M., and Sclabassi, R. J.: An Adaptive Training Algorithm for Back-Propagation Neural Networks. IEEE Trans on systems, man and cybernetics, Vol. 25, No. 3, pp. 512–514 (1995).

    Article  Google Scholar 

  3. Castellano, G., Fanelli, A. M., and Pelillo, M.: An Iterative Pruning Algorithm for Feedforward Neural Networks. IEEE Trans Neural Networks, Vol. 8, No. 3, pp. 519–531 (1997).

    Article  Google Scholar 

  4. Heskes, T.: On’ Natural’ Learning and Pruning in Multilayered Perceptrons. Neural Computation, Vol. 12, No. 4, pp. 881–901 (2000).

    Article  Google Scholar 

  5. Ponnapalli, P. V. S., Ho, K. C., and Thomson, M.: A Formal Selection and Pruning Algorithm for Feedforward Artificial Neural Network Optimization. IEEE Trans Neural Networks, Vol. 10, No. 4, pp. 964–968 (1999).

    Article  Google Scholar 

  6. Abdel-Wahhab, O., and Sid-Ahmed, M. A.: A new scheme for training feed-forward neural networks. Pattern Recognition, Vol. 30, No. 3, pp. 519–524 (1997).

    Article  Google Scholar 

  7. Karras, D. A., and Perantonis, S. J.: An Efficient Constrained Training Algorithm for Feedforward Networks. IEEE Trans Neural Networks, Vol. 6, No. 6, pp. 1420–1434 (1995).

    Article  Google Scholar 

  8. McLoone, S., Brown, M. D., Irwin, G., and Lightbody, G.: A Hybrid Linear/Nonlinear Training Algorithm for Feedforward Neural Networks. IEEE Trans Neural Networks, Vol. 9, No. 4, pp. 669–684 (1998).

    Article  Google Scholar 

  9. Tsukamoto, N., and Takahashi, H.: Generalization Performance: SVM vs. Complexity-Regularization. Technical Report of IEICE, Vol. 101, No. 154, pp. 49–55 (2001).

    Google Scholar 

  10. Fine, T. L. and Mukherjee, S.: Parameter Convergence and Learning Curves for Neural Networks. Neural Computation, Vol. 11, No. 3, pp. 747–769 (1999).

    Article  Google Scholar 

  11. Zhu, J., and Hastie, T.: Kernel Logistic Regression and the Import Vector Machine (2001). For online version: http://www.stanford.edu/ jzhu/research/nips01.ps.

  12. Lee, Y., Lin, Y., and Wahba, G.: Multicategory Support Vector Machines, Proceedings of the 33rd Symposium on the Interface, Costa Mesa, CA (2001).

    Google Scholar 

  13. Campbell, C., and Cristianini, N.: Simple Learning Algorithms for Training Support Vector Machine, Technical report, University of Bristol (1998).

    Google Scholar 

  14. Haykin, S.: Neural Networks-A comprehensive foundation, 2nd ed., Prentice Hall International, Inc (1999).

    Google Scholar 

  15. Cristanni, N, and Shawe-Taylor, J.: An Introduction to Support Vector Machines and other kernel-based learning methods (pp. 93–121), Cambridge University Press (2000).

    Google Scholar 

  16. Buzzi, C., and Grippo, L.: Convergent Decomposition Techniques for Training RBF Neural Networks, Neural Computation, Vol. 13, No. 8, pp. 1891–1920 (2001).

    Article  MATH  Google Scholar 

  17. Cohn, D.: Separating Formal Bounds from Practical Performance in Learning Systems (1992). For online version: http://citeseer.nj.nec.com/cohn92separating.html.

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2002 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Debnath, R., Takahashi, H. (2002). Learning Capability: Classical RBF Network vs. SVM with Gaussian Kernel. In: Hendtlass, T., Ali, M. (eds) Developments in Applied Artificial Intelligence. IEA/AIE 2002. Lecture Notes in Computer Science(), vol 2358. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-48035-8_29

Download citation

  • DOI: https://doi.org/10.1007/3-540-48035-8_29

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-43781-9

  • Online ISBN: 978-3-540-48035-8

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics