skip to main content
10.1145/2911451.2911548acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

Collaborative Filtering Bandits

Published:07 July 2016Publication History

ABSTRACT

Classical collaborative filtering, and content-based filtering methods try to learn a static recommendation model given training data. These approaches are far from ideal in highly dynamic recommendation domains such as news recommendation and computational advertisement, where the set of items and users is very fluid. In this work, we investigate an adaptive clustering technique for content recommendation based on exploration-exploitation strategies in contextual multi-armed bandit settings. Our algorithm takes into account the collaborative effects that arise due to the interaction of the users with the items, by dynamically grouping users based on the items under consideration and, at the same time, grouping items based on the similarity of the clusterings induced over the users. The resulting algorithm thus takes advantage of preference patterns in the data in a way akin to collaborative filtering methods. We provide an empirical analysis on medium-size real-world datasets, showing scalability and increased prediction performance (as measured by click-through rate) over state-of-the-art methods for clustering bandits. We also provide a regret analysis within a standard linear stochastic noise setting.

References

  1. C. Gentile, S. Li, and G. Zappella. Online clustering of bandits. In ICML, 2014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. N. Korda, B. Szorenyi, and S. Li, Distributed Clustering of Linear Bandits in Peer to Peer Networks. In ICML, 2016.Google ScholarGoogle Scholar
  3. Y. Abbasi-Yadkori, D. Pál, and C. Szepesvári. Improved algorithms for linear stochastic bandits. 2011.Google ScholarGoogle Scholar
  4. J.-Y. Audibert, R. Munos, and C. Szepesvári. Exploration-exploitation tradeoff using variance estimates in multi-armed bandits. Theoretical Computer Science, 410(19):1876--1902, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. P. Auer. Using confidence bounds for exploration-exploitation trade-offs. Journal of Machine Learning Research, 3:397--422, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time analysis of the multiarmed bandit problem. Machine Learning, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. M. G. Azar, A. Lazaric, and E. Brunskill. Sequential transfer in multi-armed bandit with finite set of models. In NIPS, pages 2220--2228, 2013.Google ScholarGoogle Scholar
  8. G. Bresler, G. Chen, and S. D. A latent source model for online collaborative filtering. In NIPS. MIT Press, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. E. Brunskill and L. Li. Sample complexity of multi-task reinforcement learning. In UAI, 2013.Google ScholarGoogle Scholar
  10. W. Cao, J. Li, Y. Tao, and Z. Li. On top-k selection in multi-armed bandits and hidden bipartite graphs. In NIPS, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. G. Cavallanti, N. Cesa-Bianchi, and C. Gentile. Tracking the best hyperplane with a simple budget perceptron. Machine Learning, 69/2:143--167, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. W. Chu, L. Li, L. Reyzin, and R. E. Schapire. Contextual bandits with linear payoff functions. In AISTATS, 2011.Google ScholarGoogle Scholar
  13. K. Crammer and C. Gentile. Multiclass classification with bandit feedback using adaptive regularization. In ICML, 2011.Google ScholarGoogle Scholar
  14. I. S. Dhillon. Co-clustering documents and words using bipartite spectral graph partitioning. In 7th KDD, pages 269--274. ACM, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. I. S. Dhillon, S. Mallela, and D. S. Modha. Information-theoretic co-clustering. In 9th KDD, pages 89--98, New York, NY, USA, 2003. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. J. Djolonga, A. Krause, and V. Cevher. High-dimensional gaussian process bandits. In NIPS, pages 1025--1033, 2013.Google ScholarGoogle Scholar
  17. M. Dudik, D. Erhan, J. Langford, and L. Li. Sample-efficient nonstationary-policy evaluation for contextual bandits. In UAI, 2012.Google ScholarGoogle Scholar
  18. T. George and S. Merugu. A scalable collaborative filtering framework based on co-clustering. In 5th ICDM, pages 625--628. IEEE Computer Society, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. F. Hao, S. Li, G. Min, H. Kim, S. Yau, and L. Yang,An Efficient Approach to Generating Location-Sensitive Recommendations in Ad-hoc Social Network Environments. IEEE Transactions on Services Computing, 8:3, pp. 520--533, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  20. J. Kawale, H. Bui, B. Kveton, L. Thanh, and S. Chawla. Efficient thompson sampling for online matrix-factorization recommendation. In NIPS, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. A. Krause and C. Ong. Contextual gaussian process bandit optimization. In 25th NIPS, 2011.Google ScholarGoogle Scholar
  22. B. Kveton, C. Szepesvari, Z. Wen, and A. Ashkan. Cascading bandits: Learning to rank in the cascade model. In ICML, 2015.Google ScholarGoogle Scholar
  23. L. Li, W. Chu, J. Langford, and R. Schapire. A contextual-bandit approach to personalized news article recommendation. In WWW, pages 661--670, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. S. Li, F. Hao, M. Li, and H.C. Kim.Medicine Rating Prediction and Recommendation in Mobile Social Networks.In International Conference on Green, Pervasive and Cloud Computing, 2013.Google ScholarGoogle Scholar
  25. L. Tang, Y. Jiang, L. Li, and T. Li. Ensemble contextual bandits for personalized recommendation. In RecSys, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. S. Li, C. Gentile, and A. Karatzoglou. Graph clustering bandits for recommendation. arXiv:1605.00596.Google ScholarGoogle Scholar
  27. O. Maillard and S. Mannor. Latent bandits. In ICML, 2014.Google ScholarGoogle Scholar
  28. E. Moroshko, N. Vaits, and K. Crammer. Second-order non-stationary online learning for regression. Journal of Machine Learning Research, 16:1481--1517, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. T. T. Nguyen and H. W. Lauw. Dynamic clustering of contextual multi-armed bandits. In 23rd CIKM, pages 1959--1962. ACM, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Y. Seldin, P. Auer, F. Laviolette, J. Shawe-Taylor, and R. Ortner. Pac-bayesian analysis of contextual bandits. In NIPS, pages 1683--1691, 2011.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. I. Sutskever, R. Salakhutdinov, and J. Tenenbaum. Modelling relational data using bayesian clustered tensor factorization. In NIPS, pages 1821--1828. MIT Press, 2009.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. L. Tang, Y. Jiang, L. Li, C. Zeng, and T. Li. Personalized recommendation via parameter-free contextual bandits. In SIGIR. ACM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. K. Verstrepen and B. Goethals. Unifying nearest neighbors collaborative filtering. In RecSys, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Y. Yue, S. A. Hong, and C. Guestrin. Hierarchical exploration for accelerating contextual bandits. In ICML, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Collaborative Filtering Bandits

                    Recommendations

                    Comments

                    Login options

                    Check if you have access through your login credentials or your institution to get full access on this article.

                    Sign in
                    • Published in

                      cover image ACM Conferences
                      SIGIR '16: Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval
                      July 2016
                      1296 pages
                      ISBN:9781450340694
                      DOI:10.1145/2911451

                      Copyright © 2016 ACM

                      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                      Publisher

                      Association for Computing Machinery

                      New York, NY, United States

                      Publication History

                      • Published: 7 July 2016

                      Permissions

                      Request permissions about this article.

                      Request Permissions

                      Check for updates

                      Qualifiers

                      • research-article

                      Acceptance Rates

                      SIGIR '16 Paper Acceptance Rate62of341submissions,18%Overall Acceptance Rate792of3,983submissions,20%

                    PDF Format

                    View or Download as a PDF file.

                    PDF

                    eReader

                    View online with eReader.

                    eReader