skip to main content
10.1145/2396761.2396793acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

An analysis of how ensembles of collective classifiers improve predictions in graphs

Published:29 October 2012Publication History

ABSTRACT

We present a theoretical analysis framework that shows how ensembles of collective classifiers can improve predictions for graph data. We show how collective ensemble classification reduces errors due to variance in learning and more interestingly inference. We also present an empirical framework that includes various ensemble techniques for classifying relational data using collective inference. The methods span single- and multiple-graph network approaches, and are tested on both synthetic and real world classification tasks. Our experimental results, supported by our theoretical justifications, confirm that ensemble algorithms that explicitly focus on both learning and inference processes and aim at reducing errors associated with both, are the best performers.

References

  1. A. V. Assche, C. Vens, H. Blockeel, and S. Dzeroski. A random forest approach to relational learning. In ICML'04 Workshop on SRL and its Connections.Google ScholarGoogle Scholar
  2. R. S. Y. F. P. Bartlett and W. Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. In ICML'97.Google ScholarGoogle Scholar
  3. L. Breiman. Bagging predictors. MLJ'96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. P. Domingos. A unified bias-variance decomposition for zero-one and squared loss. In AAAI'00. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. H. Eldardiry and J. Neville. Across-model collective ensemble classification. In AAAI'11.Google ScholarGoogle Scholar
  6. H. Eldardiry and J. Neville. A resampling technique for relational data graphs. In SNA-SIGKDD'08.Google ScholarGoogle Scholar
  7. A. Fast and D. Jensen. Why stacked models perform effective collective classification. In ICDM'08. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. J. Friedman. On bias, variance, 0/1-loss, and the curse-of-dimensionality. DMKD'97. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. S. Geman, E. Bienenstock, and R. Doursat. Neural networks and the bias/variance dilemma. NC'92. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. A. HeB and N.Kushmerick. Iterative ensemble classification for relational data: a case study of semantic web services. In ECML'04.Google ScholarGoogle Scholar
  11. G. James. Variance and bias for general loss functions. MLJ'03. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Z. Kou and W. W. Cohen. Stacked graphical models for effecient inference for markov random fields. In SDM'07.Google ScholarGoogle Scholar
  13. S. Natarajan, T.Khot, K. Kersting, B. Gutmann, and J. Shavlik. Gradient-based boosting for statistical relational learning: The relational dependecy network case. MLJ'12. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. J. Neville and D. Jensen. A bias/variance decomposition for models using collective inference. MLJ'08. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. J. Neville and D. Jensen. Leveraging relational autocorrelation with latent group models. In ICDM'05. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. J. Neville and D. Jensen. Relational dependency networks. JMLR'07. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. C. Preisach and L. Schmidt-Thieme. Ensembles of relational classifiers. KIS'08. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J. Quinlan. Bagging, boosting and c4.5. In AAAI'96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. K. Tumer and J. Ghosh. Analysis of decision boundaries in linearly combined neural classifiers. Pattern Recognition'96.Google ScholarGoogle Scholar
  20. R. Xiang and J. Neville. Understanding propagation error and its effect on collective classification. In ICDM'11. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Y.Freund and R.E.Schapire. Experiments with a new boosting algorithm. In ICML'96.Google ScholarGoogle Scholar

Index Terms

  1. An analysis of how ensembles of collective classifiers improve predictions in graphs

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CIKM '12: Proceedings of the 21st ACM international conference on Information and knowledge management
      October 2012
      2840 pages
      ISBN:9781450311564
      DOI:10.1145/2396761

      Copyright © 2012 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 29 October 2012

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate1,861of8,427submissions,22%

      Upcoming Conference

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader