Abstract
In this paper, we introduce a new mining algorithm to improve the classification accuracy rates aiming at the deficiency of the typical K-Dependence Bayes (KDB) model which ignores the topology changing as the result of inputing of test instances. Under this condition, we put forward an algorithm called Base K-Dependence Bayes (B-KDB) which consists of a Label-based K-Dependence Bayes (L-KDB) algorithm and a Instance-based K-Dependence Bayes (I-KDB) algorithm. The I-KDB algorithm is used to build I-KDB model by instances to be tested and it can deal with the problem of test instances topology keep on changing. However I-KDB model is extraordinarily sensitive to data and it may suffer from overfitting and the effect of noisy instances, therefore L-KDB algorithm is designed as complement. After combing these two algorithms into B-KDB, we built B-KDB model and tested the performance against the KDB model in classification accuracy, precision, sensitivity-specificity analysis with 10-fold cross validation on 55 real benchmark datasets from University of California Irvine (UCI) machine learning repository. The experimental result, which shows the classification accuracy of our model twice as much as KDB, indicates our algorithm efficient and proves our idea of improving the KDB algorithm classification accuracy feasible.
Y. Wang and J. Guo—These authors contributed equally to this work.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Bielza, C., Larranaga, P.: Discrete Bayesian network classifiers: a survey. ACM Comput. Surv. 47, 1–43 (2014)
Sloin, A., Wiesel, A.: Proper quaternion Gaussian graphical models. IEEE. Trans. Signal Process. 62, 5487–5496 (2014)
Cheng, J., Greiner, R., Kelly, J., Bell, D., Liu, W.: Learning Bayesian networks from data: an information-theory based approach. Artif. Intell. 137, 43–90 (2002)
Cooper, G.F.: The computational complexity of probabilistic inference using Bayesian belief networks. Artif. Intell. 42, 393–405 (1990)
Minsky, M.: Steps toward artificial intelligence. Proc. IRE 49, 8–30 (1961)
Friedman, N., Geiger, D., Goldszmidt, M.: Bayesian network classifiers. Mach. Learn. 29, 131–163 (1997)
Webb, G.I., Boughton, J., Wang, Z.: Not so Naïve Bayes: aggregating one-dependence estimators. Mach. Learn. 58, 5–24 (2005)
Sahami, M.: Learning limited dependence Bayesian classifiers. In: Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD 1996), Portland, OR, USA, pp. 335–338, 2–4 August 1996
Zhang, H., Sheng, S.: Learning weighted Naïve Bayes with accurate ranking. In: The 4th IEEE International Conference on Data Ming. IEEE Computer Society, Chicago (2004)
Asuncion, A., Newman, D.J.: UCI Machine Learning Repository. University of California, Department of Information and Computer Science, Irvine, CA (2007). http://www.ics.uci.edu/mlearn/MLRepository.html
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Wang, L., Xie, Y., Zhou, H., Wang, Y., Guo, J. (2016). Learning Based K-Dependence Bayesian Classifiers. In: Sun, X., Liu, A., Chao, HC., Bertino, E. (eds) Cloud Computing and Security. ICCCS 2016. Lecture Notes in Computer Science(), vol 10040. Springer, Cham. https://doi.org/10.1007/978-3-319-48674-1_49
Download citation
DOI: https://doi.org/10.1007/978-3-319-48674-1_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-48673-4
Online ISBN: 978-3-319-48674-1
eBook Packages: Computer ScienceComputer Science (R0)