Skip to main content

Selecting Features by Learning Markov Blankets

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4692))

Abstract

In this paper I propose a novel feature selection technique based on Bayesian networks. The main idea is to exploit the conditional independencies entailed by Bayesian networks in order to discard features that are not directly relevant for classification tasks. An algorithm for learning Bayesian networks and its use in feature selection are illustrated. The advantages of this algorithm with respect to other ones are then discussed. Finally, experimental results are offered which confirm the reliability of the algorithm.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. John Wiley & Sons, New York (NY) (2001)

    MATH  Google Scholar 

  2. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. Journal of Machine Learning Research 3, 1157–1182 (2003)

    Article  MATH  Google Scholar 

  3. John, G.H., Kohavi, R., Pfleger, K.: Irrelevant features and the subset selection problem. In: Cohen, W.W., Hirsh, H. (eds.) Machine Learning: Proceedings of the Eleventh International Conference, pp. 121–129. Morgan Kaufmann, San Francisco (1994)

    Google Scholar 

  4. Kohavi, R., John, G.H.: Wrappers for feature subset selection. Artificial Intelligence 97, 273–324 (1997)

    Article  MATH  Google Scholar 

  5. Koller, D., Sahami, M.: Toward optimal feature selection. In: Saitta, L. (ed.) Machine Learning: Proceedings of the Thirteenth International Conference, pp. 284–292. Morgan Kaufmann, San Francisco (1996)

    Google Scholar 

  6. Hruschka, E.R.J., Hruschka, E.R., Ebecken, N.F.F.: Feature selection by Bayesian networks. In: Tawfik, A.Y., Goodwin, S.D. (eds.) Advances in Artificial Intelligence, pp. 370–379. Springer, Heidelberg (2004)

    Google Scholar 

  7. Pearl, J.: Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann, San Francisco (CA) (1988)

    Google Scholar 

  8. Neapolitan, R.E.: Learning Bayesian Networks. Prentice Hall, Upper Saddle River (NJ) (2004)

    Google Scholar 

  9. Russell, S., Norvig, P.: Artificial Intelligence, 2nd edn. Prentice Hall, Upper Saddle River (NJ) (2003)

    Google Scholar 

  10. Heckerman, D.: A tutorial on learning with Bayesian networks. Technical Report MSR-TR-95-06, Microsoft Research, Redmond (WA) (1996)

    Google Scholar 

  11. Witten, I.H., Frank, E.: Data Mining, 2nd edn. Morgan Kaufmann, San Francisco (CA) (2005)

    MATH  Google Scholar 

  12. Rissanen, J.: Stochastic complexity. Journal of the Royal Statistical Society.Series B 49, 223–239 (1987)

    MATH  Google Scholar 

  13. Cooper, G.F., Herskovits, E.: A Bayesian method for the induction of probabilistic networks from data. Machine Learning 9, 309–347 (1992)

    MATH  Google Scholar 

  14. Newman, D., Hettich, S., Blake, C., Merz, C.: UCI repository of machine learning databases (1998)

    Google Scholar 

  15. Pollastro, P., Rampone, S.: Homo sapiens splice sites dataset (2003)

    Google Scholar 

  16. Domingos, P., Pazzani, M.: On the optimality of the simple Bayesian classifier under zero-one loss. Machine Learning 29, 103–130 (1997)

    Article  MATH  Google Scholar 

  17. Quinlan, J.R.: C4.5 Morgan Kaufmann, San Mateo (CA) (1993)

    Google Scholar 

  18. Aha, D.W., Kibler, D., Albert, M.K.: Instance-based learning algorithms. Machine Learning 6, 37–66 (1991)

    Google Scholar 

  19. Hall, M.A.: Correlation-based feature selection for discrete and numeric class machine learning. In: Proceedings of the Seventeenth International Conference on Machine Learning (ICML 2000), pp. 359–366. Morgan Kaufmann, San Francisco (2000)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Bruno Apolloni Robert J. Howlett Lakhmi Jain

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Freno, A. (2007). Selecting Features by Learning Markov Blankets. In: Apolloni, B., Howlett, R.J., Jain, L. (eds) Knowledge-Based Intelligent Information and Engineering Systems. KES 2007. Lecture Notes in Computer Science(), vol 4692. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74819-9_9

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-74819-9_9

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-74817-5

  • Online ISBN: 978-3-540-74819-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics