Abstract
This paper proposed a novel hybrid probabilistic network, which is a good tradeoff between the model complexity and learnability in practice. It relaxes the conditional independence assumptions of Naive Bayes while still permitting efficient inference and learning. Experimental studies on a set of natural domains prove its clear advantages with respect to the generalization ability.
Similar content being viewed by others
References
Lin W M, Lin C H, Tasy M X. Transformer-fault diagnosis by integrating field data and standard codes with training enhancible adaptive probabilistic network. In: IEE Proceedings of Generation, Transmission and Distribution. 2005, 152: 335–341
Tseng C L, Chen Y H, Xu Y Y, et al. A self-growing probabilistic decision-based neural network with automatic data clustering. Neurocomputing, 2004, 61: 21–38
Malgorzata, Sethi A S. Probabilistic fault localization in communication systems using belief networks Steinder. IEEE/ACM Transactions on Networking, 2004, 12: 809–822
Specht D F. Probabilistic neural networks. Neural Networks, 1990, 3: 109–118
Kononenko I. Semi-naive Bayesian classifier. Proceedings of sixth European working session on learning, 1991, 206–219
Langley P, Iba W, Thompson K. An analysis of Bayesian classifiers. In: Proceedings of AAAI-92. 1992, 223–228
Friedman N, Geiger D, Goldszmidt M. Bayesian Network Classifiers. Machine Learning, 1997, 29: 131–163
Pazzani M J, Keogh E J. Learning Augmented Bayesian Classifiers: A Comparison of Distribution-Based and Classification-Based Approaches. In: Proceedings of the Seventh International Workshop on Artificial Intelligence and Statistics, 1999, 225–230
Hendler J. Developing hybrid symbolic/connectionist models, Advances in Connectionist and Neural Computation Theory. 1991, 165–179
Sun R, Peterson T. A Hybrid Model for Learning Sequential Navigation. In: Proceedings of the 11th International Parallel Processing Symposium, 1995, 205–217
Friedman N, Goldszmidt M, Thomas J L. Bayesian Network Classification with Continuous Attributes: Getting the Best of Both Discretization and Parametric Fitting. In: Proceedings of the International Conference on Machine Learning. 1998, 179–187
Dougherty J. Supervised and unsupervied discretization of coninuous features. In: Proceedings of the 12th International Conference on Machine Learning, 1995, 194–201
Chow C K, Liu C N. Approximating Discrete Probability Distributions with Dependence Trees. IEEE Transactions on Information Theory, 1968, 14: 462–267
Specht D F. Probabilstic neural networks, Neural Networks, 1990, 3: 109–118
Hunter A. Feature Selection Using Probabilistic Neural Networks. Neural Computing and Applications, 2000, 9: 124–132
Fayyad U M, Irani K B. Multi-interval discretization of continuous valued attributes for classification learning. In: Proceedings of the 13th International Conference on Artificial Intelligence. 1993, 1022–1027
Dougherty J, Kohavi R, Sahami M. Supervised and unsupervised discretization of continuous features. In: Proceedings of the International Conference on Mathematical Linguistics. 1995, 194–202
Silverman B W. Density Estimation for Statistics and Data Analysis. Monographs on Statistics and Applied Probability, 1986.
George H. Estimating Continuous Distributions in Bayesian Classifiers. In: Proceedings of the 11th International Conference on Uncertainty in Artificial Intelligence, 1995, 338–356
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Wang, L., Wang, X. & Li, X. Inference and learning in hybrid probabilistic network. Front. Comput. Sc. China 1, 429–435 (2007). https://doi.org/10.1007/s11704-007-0041-0
Received:
Accepted:
Issue Date:
DOI: https://doi.org/10.1007/s11704-007-0041-0