skip to main content
10.1145/2020408.2020504acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Localized factor models for multi-context recommendation

Published:21 August 2011Publication History

ABSTRACT

Combining correlated information from multiple contexts can significantly improve predictive accuracy in recommender problems. Such information from multiple contexts is often available in the form of several incomplete matrices spanning a set of entities like users, items, features, and so on. Existing methods simultaneously factorize these matrices by sharing a single set of factors for entities across all contexts. We show that such a strategy may introduce significant bias in estimates and propose a new model that ameliorates this issue by positing local, context-specific factors for entities. To avoid over-fitting in contexts with sparse data, the local factors are connected through a shared global model. This sharing of parameters allows information to flow across contexts through multivariate regressions among local factors, instead of enforcing exactly the same factors for an entity, everywhere. Model fitting is done in an EM framework, we show that the E-step can be fitted through a fast multi-resolution Kalman filter algorithm that ensures scalability. Experiments on benchmark and real-world Yahoo! datasets clearly illustrate the usefulness of our approach. Our model significantly improves predictive accuracy, especially in cold-start scenarios.

References

  1. D. Agarwal and B.-C. Chen. Regression-based latent factor models. In KDD, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. D. Agarwal and B.-C. Chen, et al. Online models for content optimization. In NIPS, 2008.Google ScholarGoogle Scholar
  3. R. Ando and T. Zhang. A high-performance semi-supervised learning method for text chunking. In ACL, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. A. Argyriou, T. Evgeniou, and M. Pontil. Multi-task feature learning. In NIPS, 2007.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Argyriou, C. Micchelli, M. Pontil, and Y. Ying. A spectral regularization framework for multi-task structure learning. NIPS, 2008.Google ScholarGoogle Scholar
  6. S. Bickel, M. Brückner, and T. Scheffer. Discriminative learning for differing training and test distributions. In ICML, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. D. Blackwell. Conditional expectation and unbiased sequential estimation. Annals of Math. Stat., 1947.Google ScholarGoogle ScholarCross RefCross Ref
  8. J. Blitzer, K. Crammer, A. Kulesza, F. Pereira, and J. Wortman. Learning bounds for domain adaptation. NIPS, 2008.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. J. Blitzer, R. McDonald, and F. Pereira. Domain adaptation with structural correspondence learning. In EMNLP, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training. In COLT, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. E. Bonilla, K. Chai, and C. Williams. Multi-task gaussian process prediction. NIPS, 2007.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. R. Carroll, D. Ruppert, L. Stefanski, and C. Crainiceanu. Measurement Error in Nonlinear Models. Chapman and Hall, 2008.Google ScholarGoogle Scholar
  13. K. C. Chou, A. S. Willsky, and R. Nikoukhah. Multiscale systems, Kalman filters, and Ricatti equations. IEEE Trans. on Automatic Control, 1994.Google ScholarGoogle Scholar
  14. W. Dai, G. Xue, Q. Yang, and Y. Yu. Co-clustering based classification for out-of-domain documents. In KDD, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. W. Dai, Q. Yang, G. Xue, and Y. Yu. Boosting for transfer learning. In ICML, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. H. Daumé. Cross-task knowledge-constrained self training. In EMNLP, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. H. Daume III and D. Marcu. Domain adaptation for statistical classifiers. Journal of Artificial Intelligence Research, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. J. Davis and P. Domingos. Deep transfer via second-order markov logic. In AAAI Workshop on Transfer Learning for Complex Tasks, 2008.Google ScholarGoogle Scholar
  19. A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. J. of the Royal Statistical Society, Series B, 1977.Google ScholarGoogle Scholar
  20. T. Evgeniou and M. Pontil. Regularized multi-task learning. In KDD, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. J. L. Herlocker, J. A. Konstan, A. Borchers, and J. Riedl. An algorithmic framework for performing collaborative filtering. In SIGIR, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. H. Huang and N. Cressie. Fast spatial prediction of global processes from satellite data. SIAM J. on Scientific Computing, 2000.Google ScholarGoogle Scholar
  23. H. Huang and N. Cressie. Multiscale graphical modeling in space: Applications to command and control. In Spatial Statistics Workshop, 2000.Google ScholarGoogle Scholar
  24. J. Huang, A. Smola, A. Gretton, K. Borgwardt, and B. Scholkopf. Correcting sample selection bias by unlabeled data. NIPS, 2007.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. T. Jaakkola and M. Jordan. Bayesian logistic regression: a variational approach. Statistics and Computing, 2000.Google ScholarGoogle Scholar
  26. J. Jiang and C. Zhai. Instance weighting for domain adaptation in NLP. In ACL, 2007.Google ScholarGoogle Scholar
  27. H. Kaji and Y. Morimoto. Unsupervised word sense disambiguation using bilingual comparable corpora. In COLIN, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Y. Koren, R. Bell, and C. Volinsky. Matrix factorization techniques for recommender systems. Computer, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. N. Kushmerick. Learning to remove internet advertisements. In Int. Conf. Autonomous Agents, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. N. Lawrence and J. Platt. Learning to learn with the informative vector machine. In ICML, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. S. Lee, V. Chatalbashev, D. Vickrey, and D. Koller. Learning a meta-level prior for feature relevance from multiple related tasks. In ICML, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. X. Liao, Y. Xue, and L. Carin. Logistic regression with an auxiliary data source. In ICML, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. X. Ling, W. Dai, G.-R. Xue, Q. Yang, and Y. Yu. Spectral domain-transfer learning. In KDD, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. P. Massa and P. Avesani. Trust metrics in recommender systems. In Computing with Social Trust, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  35. L. Mihalkova, T. Huynh, and R. Mooney. Mapping and revising markov logic networks for transfer learning. In AAAI, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. L. Mihalkova and R. Mooney. Transfer learning by mapping with minimal target data. In AAAI Workshop on Transfer Learning for Complex Tasks, 2008.Google ScholarGoogle Scholar
  37. S. J. Pan and Q. Yang. A survey on transfer learning. Technical Report HKUST-CS08-08, Hong Kong University of Science and Technology, 2008.Google ScholarGoogle Scholar
  38. R. Raina, A. Battle, H. Lee, B. Packer, and A. Ng. Self-taught learning: Transfer learning from unlabeled data. In ICML, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. A. Schwaighofer, V. Tresp, and K. Yu. Learning Gaussian process kernels via hierarchical Bayes. NIPS, 2005.Google ScholarGoogle Scholar
  40. A. P. Singh and G. J. Gordon. Relational learning via collective matrix factorization. In KDD, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. D. H. Stern, R. Herbrich, and T. Graepel. Matchbox: large scale online bayesian recommendations. In WWW, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. X. Su and T. M. Khoshgoftaar. A survey of collaborative filtering techniques. Adv. in Artif. Intell., 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. M. Sugiyama, S. Nakajima, H. Kashima, P. von Bunau, and M. Kawanabe. Direct importance estimation with model selection and its application to covariate shift adaptation. NIPS, 2008.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. K. Yu and V. Tresp. Learning to learn and collaborative filtering. In In NIPS Workshop on Inductive Transfer: 10 Years Later, 2005.Google ScholarGoogle Scholar
  45. S. Yu, K. Yu, V. Tresp, and H.-P. Kriegel. Collaborative ordinal regression. In ICML, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Localized factor models for multi-context recommendation

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in
            • Published in

              cover image ACM Conferences
              KDD '11: Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining
              August 2011
              1446 pages
              ISBN:9781450308137
              DOI:10.1145/2020408

              Copyright © 2011 ACM

              Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

              Publisher

              Association for Computing Machinery

              New York, NY, United States

              Publication History

              • Published: 21 August 2011

              Permissions

              Request permissions about this article.

              Request Permissions

              Check for updates

              Qualifiers

              • research-article

              Acceptance Rates

              Overall Acceptance Rate1,133of8,635submissions,13%

              Upcoming Conference

              KDD '24

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader