Abstract
Deep learning methods aims at learning features automatically at multiple levels that allow the system to learn complex functions mapping the input to the output directly from data. This ability to automatically learn powerful features will become increasingly important as the amount of data and range of applications to machine learning methods continues to grow. In this context we propose a deep architecture model using Support Vector Machine (SVM) which has inherent ability to select data points important for classification with good generalization capabilities. Since SVM can effectively discriminate features, we used support vectors with kernel as non-linear discriminant features for classification. By stacking SVMs in to multiple layers, we can obtain deep features without extra feature engineering steps and get robust recognition accuracy. Experimental results show that the proposed method improves generalization performance on Wisconsin Breast Cancer dataset.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Arel, I., Rose, D.C., Karnowski, T.P.: Deep machine learning-A new frontier in artificial intelligence research. IEEE Computational Intelligence Magazine 5, 13–18 (2010)
Bengio, Y.: Learning deep architectures for AI. Foundations and Trends in Machine Learning 2 (2009)
Utgoff, P.E., David, J.S.: Many-layered learning. Neural Computation 14, 2497–2529 (2002)
Tesauro, G.: Practical issues in temporal difference learning. Reinforcement Learning, pp. 33–53. Springer US (1992)
LeCun, Y., et al.: Backpropagation applied to handwritten zip code recognition. Neural computation 1, 541–551 (1989)
Fukushima, K.: Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biological cybernetics 36, 193–202 (1980)
Smolensky, P.: Information processing in dynamical systems: Foundations of harmony theory (1986)
Hinton, G.E., Osindero, S., Teh, Y.-W.: A fast learning algorithm for deep belief nets. Neural Computation 18, 1527–1554 (2006)
Hinton, G.E.: A practical guide to training restricted Boltzmann machines. Momentum 9 (2010)
Cortes, C., Vapnik, V.: Support-vector networks. Machine Learning 20, 273–297 (1995)
Bishop, C.M.: Pattern recognition and machine learning. Springer, New York (2006)
Bennett, K.P., Mangasarian, O.L.: Robust linear programming discrimination of two linearly inseparable sets. Optimization Methods and Software 1, 23–34 (1992)
Vincent, P., et al.: Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. The Journal of Machine Learning Research 9999, 3371–3408 (2010)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kim, S., Kavuri, S., Lee, M. (2013). Deep Network with Support Vector Machines. In: Lee, M., Hirose, A., Hou, ZG., Kil, R.M. (eds) Neural Information Processing. ICONIP 2013. Lecture Notes in Computer Science, vol 8226. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-42054-2_57
Download citation
DOI: https://doi.org/10.1007/978-3-642-42054-2_57
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-42053-5
Online ISBN: 978-3-642-42054-2
eBook Packages: Computer ScienceComputer Science (R0)