Skip to main content
Log in

From Gaussian kernel density estimation to kernel methods

  • Original Article
  • Published:
International Journal of Machine Learning and Cybernetics Aims and scope Submit manuscript

Abstract

This paper explores how a kind of probabilistic systems, namely, Gaussian kernel density estimation (GKDE), can be used to interpret several classical kernel methods, including the well-known support vector machine (SVM), support vector regression (SVR), one-class kernel classifier, i.e., support vector data description (SVDD) or equivalently minimal enclosing ball (MEB), and the fuzzy systems (FS). For the SVM, we reveal that the classical SVM with Gaussian density kernel attempts to find a noisy GKDE based Bayesian classifier with equal prior probabilities for each class. For the SVR, the classification based ε-SVR attempts to obtain two noisy GKDEs for each class in the constructed binary classification dataset, and the decision boundary just corresponds to the mapping function of the original regression problem. For the MEB or SVDD, we reveal the equivalence between it and the integrated-squared-errors (ISE) criterion based GKDE and by using this equivalence a MEB based classifier with privacy-preserving function is proposed for one kind of classification tasks where the datasets contain privacy-preserving clouds. For the FS, we show that the GKDE for a regression dataset is equivalent to the construction of a zero-order Takagi–Sugeno–Kang (TSK) fuzzy system based on the same dataset. Our extensive experiments confirm the obtained conclusions and demonstrated the effectiveness of the proposed new machine learning and modeling methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Vapnik V (1995) The nature of statistical learning theory. Springer, New York

    MATH  Google Scholar 

  2. Tax D, Duin R (2004) Support vector data description. Mach Learn 54(1):45–66

    Article  MATH  Google Scholar 

  3. Tsang IW, Kwok JT, Zurada JM (2006) Generalized core vector machines. IEEE Trans Neural Netw 17(5):1126–1140

    Article  Google Scholar 

  4. Zhou SM, Gan JQ (2007) Constructing L2-SVM-based fuzzy classifiers in high-dimensional space with automatic model selection and fuzzy rule ranking. IEEE Trans Fuzzy Syst 15(3):398–409

    Article  Google Scholar 

  5. Juang CF, Hsieh CD (2009) TS-fuzzy system-based support vector regression. Fuzzy Sets Syst 160:2486–2504

    Article  MathSciNet  MATH  Google Scholar 

  6. Chiang JH, Hao PY (2004) Support vector learning mechanism for fuzzy rule-based modeling: a new approach. IEEE Trans Fuzzy Syst 12(1):1–11

    Article  Google Scholar 

  7. Chen Y, Wang JZ (2003) Support vector learning for fuzzy rule-based classification systems. IEEE Trans Fuzzy Syst 11(6):716–728

    Article  Google Scholar 

  8. Kwok JT, Tsang IW (2003) Linear dependency between epsilon and the input noise in epsilon-support vector regression. IEEE Trans Neural Netw 14(3):544–553

    Article  Google Scholar 

  9. Gan MT, Hanmandlu M, Tan AH (2005) From a Gaussian mixture model to additive fuzzy systems. IEEE Trans Fuzzy Syst 13(3):303–316

    Article  Google Scholar 

  10. Verma N, Hanmandlu M (2007) From a Gaussian mixture model to non-additive fuzzy systems. IEEE Trans Fuzzy Syst 15(5):809–826

    Article  Google Scholar 

  11. Kim J, Scott CD (2010) L2 kernel classification. IEEE Trans Pattern Anal Mach Intell 32(10):1822–1831

    Article  Google Scholar 

  12. Girolami M, He C (2003) Probability density estimation from optimally condensed data samples. IEEE Trans Pattern Anal Mach Intell 25(10):1253–1264

    Article  Google Scholar 

  13. Deng ZH, Chung FL, Wang ST (2008) FRSDE: fast reduced set density estimator using minimal enclosing ball approximation. Pattern Recogn 41(4):1363–1372

    Article  MATH  Google Scholar 

  14. Chen G, Hong X, Harris CJ (2010) Probability density estimation with tunable kernels using orthogonal forward regression. IEEE Trans Syst Man Cybern (Part B) 40(4):1101–1114

    Article  Google Scholar 

  15. Orchel M (2011) Regression based on support vector classification. http://svms.pl/SVM/Files/svcRegression_Orchel.pdf

  16. Tao Q, Cao JD, Sun DM (2002) A regression method based on the support vectors for classification. J Softw 13(5):1024–1028

    Google Scholar 

  17. Chen K, Liu L (2005) Privacy preserving data classification with rotation perturbation. In: Proceedings of IEEE ICDM’05, pp 589–592

  18. Vaidya J, Yu H, Jiang X (2008) Privacy preserving SVM classification. Knowl Inform Syst 14(2):161–178

    Article  Google Scholar 

  19. Lanckriet GRG, Ghaoui LE, Bhattacharyya C, Jordan MI (2003) A robust minimax approach to classification. J Mach Learn Res 3:555–582

    MathSciNet  MATH  Google Scholar 

  20. Stefan R (2010) SVM Classifier Estimation from Group Probabilities. In: Proceedings of the 27th international conference on machine learning (ICML 2010), Haifa, Israel

  21. Quadrianto N, Smola AJ, Caetano TS, Le QV (2009) Estimating labels from label proportions. J Mach Learn Res 10:2349–2374

    MathSciNet  MATH  Google Scholar 

  22. Zhang ZC, Chung FL, Wang ST (2012) Collaborative classification mechanism for privacy-preserving on horizontally full-partitioned data. IEEE Trans SMC (Part B) (submitted)

  23. Chung FL, Wang ST, Deng ZG, Hu DW (2004) Fuzzy Kernel Hyperball Perceptron. Appl Soft Comput 5:67–74

    Article  Google Scholar 

  24. Jang JSR, Sun CT, Mizutani E (1997) Neuro-fuzzy and soft-computing. Prentice-Hall, Upper Saddle River

    Google Scholar 

  25. Schölkopf B, Smola AJ, Williamson RC, Bartlett PL (2000) New support vector algorithms. Neural Comput 12(5):1207–1245

    Article  Google Scholar 

  26. Chung FL, Deng ZH, Wang ST (2009) From minimum enclosing ball to fast fuzzy inference system training on large datasets. IEEE Trans Fuzzy Syst 17(1):173–184

    Article  Google Scholar 

  27. Chen CY, Zhang JP, He XF, Zhou ZH (2011) Non-parametric Kernel learning with robust pairwise constraints. Int J Mach Learn Cybern. doi:10.1007/s13042-011-0048-6

  28. He Q, Wu CX (2011) Separating theorem of samples in Banach space for support vector machine learning. Int J Mach Learn Cybern 2(1):49–54

    Article  Google Scholar 

  29. Liu Z, Wu QH, Zhang Y, Chen CL (2011) Adaptive least squares support vector machines filter for hand tremor canceling in microsurgery. Int J Mach Learn Cybern 2(1):37–47

    Article  MathSciNet  Google Scholar 

  30. Wu J, Wang ST, Chung FL (2011) Positive and negative fuzzy rule system, extreme learning machine and image classification. Int J Mach Learn Cybern 2(4):261–271

    Article  Google Scholar 

  31. Wang LJ (2011) An improved multiple fuzzy NNC system based on mutual information and fuzzy integral. Int J Mach Learn Cybern 2(1):25–36

    Article  Google Scholar 

Download references

Acknowledgments

This work was supported in part by the Hong Kong Polytechnic University under Grants 1-ZV5V and G-U724, and by the National Natural Science Foundation of China under Grants 60903100, 60975027, 61170122, and by the Natural Science Foundation of Jiangsu Province under Grant BK2009067, 2011NSFJS plus its Key Grant, JiangSu 333 expert engineering grant (BRA2011142), and 2011 Postgraduate Student’s Creative Research Fund of Jiangsu Province under Grant CXZZ11-0483.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shitong Wang.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Wang, S., Deng, Z., Chung, Fl. et al. From Gaussian kernel density estimation to kernel methods. Int. J. Mach. Learn. & Cyber. 4, 119–137 (2013). https://doi.org/10.1007/s13042-012-0078-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13042-012-0078-8

Keywords

Navigation