Skip to main content

Automatic Detection of Latent Common Clusters of Groups in MultiGroup Regression

  • Conference paper
  • First Online:
Machine Learning and Data Mining in Pattern Recognition (MLDM 2016)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9729))

  • 3022 Accesses

Abstract

We present a flexible non-parametric generative model for multigroup regression that detects latent common clusters of groups. The model is founded on techniques that are now considered standard in the statistical parameter estimation literature, namely, Dirichlet process(DP) and Generalized Linear Model (GLM), and therefore, we name it “Infinite MultiGroup Generalized Linear Models” (iMG-GLM). We present two versions of the core model. First, in iMG-GLM-1, we demonstrate how the use of a DP prior on the groups while modeling the response-covariate densities via GLM, allows the model to capture latent clusters of groups by noting similar densities. The model ensures different densities for different clusters of groups in the multigroup setting. Secondly, in iMG-GLM-2, we model the posterior density of a new group using the latent densities of the clusters inferred from previous groups as prior. This spares the model from needing to memorize the entire data of previous groups. The posterior inference for iMG-GLM-1 is done using Variational Inference and that for iMG-GLM-2 using a simple Metropolis Hastings Algorithm. We demonstrate iMG-GLM’s superior accuracy in comparison to well known competing methods like Generalized Linear Mixed Model (GLMM), Random Forest, Linear Regression etc. on two real world problems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Ando, R.K., Zhang, T.: A framework for learning predictive structures from multiple tasks and unlabeled data. Journal of Machine Learning Research 6, 1817–1853 (2005)

    MathSciNet  MATH  Google Scholar 

  2. Antoniak, C.: Mixtures of dirichlet processes with applications to bayesian nonparametric problems. Annals of Statistics 2(6), 1152–1174 (1974)

    Article  MathSciNet  MATH  Google Scholar 

  3. Bakker, B., Heskes, T.: Task clustering and gating for bayesian multitask learning. Journal of Machine Learning Research 4, 83–99 (2003)

    MATH  Google Scholar 

  4. Baxter, J.: Learning internal representations. In: International Conference on Computational Learning Theory, pp. 311–320 (1995)

    Google Scholar 

  5. Baxter, J.: A model of inductive bias learning. Journal of Artificial Intelligence Research 12, 149–198 (2000)

    MathSciNet  MATH  Google Scholar 

  6. Blei, D.: Variational inference for dirichlet process mixtures. Bayesian Analysis 1, 121–144 (2006)

    Article  MathSciNet  MATH  Google Scholar 

  7. Breslow, N.E., Clayton, D.G.: Approximate inference in generalized linear mixed models. Journal of the American Statistical Association 88(421), 9–25 (1993)

    MATH  Google Scholar 

  8. Caruana, R.: Multitask learning. Machine Learning 28(1), 41–75 (1997)

    Article  MathSciNet  Google Scholar 

  9. Escobar, M.D., West, M.: Bayesian density estimation and inference using mixtures. Journal of the American Statistical Association 90, 577–588 (1994)

    Article  MathSciNet  MATH  Google Scholar 

  10. Ferguson, T.: A bayesian analysis of some nonparametric problems. Annals of Statistics 1, 209–230 (1973)

    Article  MathSciNet  MATH  Google Scholar 

  11. Hastie, T., Tibshirani, R.: Varying-coefficient models. Journal of the Royal Statistical Society. Series B (Methodological) 55(4), 757–796 (1993)

    MathSciNet  MATH  Google Scholar 

  12. IBM: Ibm spss version 20. IBM SPSS SOFTWARE (2011)

    Google Scholar 

  13. Jordan, M., Jacobs, R.: Hierarchical mixtures of experts and the EM algorithm. In: International Joint Conference on Neural Networks (1993)

    Google Scholar 

  14. Lee, Y., Nelder, J.A.: Hierarchical generalized linear models. Journal of the Royal Statistical Society. Series B (Methodological) 58(4), 619–678 (1996)

    MathSciNet  MATH  Google Scholar 

  15. Lee, Y., Nelder, J.A.: Hierarchical generalised linear models: A synthesis of generalised linear models, random-effect models and structured dispersions. Biometrika 88(4), 987–1006 (2001)

    Article  MathSciNet  MATH  Google Scholar 

  16. Lee, Y., Nelder, J.A.: Modelling and analysing correlated non-normal data. Statistical Modelling 1(1), 3–16 (2001)

    Article  MATH  Google Scholar 

  17. Lee, Y., Nelder, J.A.: Double hierarchical generalized linear models (with discussion). Journal of the Royal Statistical Society: Series C (Applied Statistics) 55(2), 139–185 (2006)

    Article  MathSciNet  MATH  Google Scholar 

  18. Neal, R.M.: Markov chain sampling methods for dirichlet process mixture models. Journal of Computational and Graphical Statistics 9(2), 249–265 (2000)

    MathSciNet  Google Scholar 

  19. Nelder, J.A., Wedderburn, R.W.M.: Generalized linear models. Journal of the Royal Statistical Society, Series A (General) 135(3), 370–384 (1972)

    Google Scholar 

  20. Robert, C., Casella, G.: Monte Carlo Statistical Methods (Springer Texts in Statistics). Springer-Verlag New York, Inc. (2005)

    Google Scholar 

  21. Sethuraman, J.: A constructive definition of dirichlet priors. Statistica Sinica 4, 639–650 (1994)

    MathSciNet  MATH  Google Scholar 

  22. Viele, K., Tong, B.: Modeling with mixtures of linear regressions. Statistics and Computing 12(4), 315–330 (2002)

    Article  MathSciNet  Google Scholar 

  23. Xue, Y., Liao, X., Carin, L.: Multi-task learning for classification with dirichlet process priors. Journal of Machine Learning Research 8, 35–63 (2007)

    Google Scholar 

  24. Yu, K., Tresp, V., Schwaighofer, A.: Learning gaussian processes from multiple tasks. International Conference on Machine Learning, pp. 1012–1019 (2005)

    Google Scholar 

  25. Zhang, J., Ghahramani, Z., Yang, Y.: Learning multiple related tasks using latent independent component analysis. Advances in Neural Information Processing Systems, pp. 1585–1592 (2005)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Minhazul Islam Sk .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Sk, M.I., Banerjee, A. (2016). Automatic Detection of Latent Common Clusters of Groups in MultiGroup Regression. In: Perner, P. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2016. Lecture Notes in Computer Science(), vol 9729. Springer, Cham. https://doi.org/10.1007/978-3-319-41920-6_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-41920-6_19

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-41919-0

  • Online ISBN: 978-3-319-41920-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics