skip to main content
10.1145/3487572.3487598acmotherconferencesArticle/Chapter ViewAbstractPublication PagesrecsysConference Proceedingsconference-collections
research-article

Addressing the cold-start problem with a two-branch architecture for fair tweet recommendation

Published:22 November 2021Publication History

ABSTRACT

In this paper we describe our solution for the RecSys Challenge 20211 focused on tweet recommendation to increase the number of user interactions. A large database with 800 million tweets produced over four weeks is used as training to predict four types of interactions: Like, Reply, Retweet and Quote. The proposed challenge is very similar to last year’s but incorporates a new fairness metric. Authors are splited into groups based on their number of followers and the final score is computed by averaging PRAUC and RCE of these groups. To satisfy this new restriction, we present a two-branch architecture that separates authors according to their total number of interactions in the dataset. In this way, authors who appear a few number of times (cold-start users) are predicted using similar users. The same is true for users with many interactions (active users). Each of the branches consists of a concatenation of four LightGBM models, one per target. They all use features we extracted from the interaction but they also use the output of the previous model. We first predict Like and use the output to predict Retweet. Then we predict Reply using Like and Retweet and so on. The users’ popularity, as well as the first and last words of the tweet text, turned out to be the best features for our method. Our solution obtained the 5th place in the final ranking and won the 2nd prize in the academic category. All the source code is available online2.

References

  1. Vito Walter Anelli, Saikishore Kalloori, Bruce Ferwerda, Luca Belli, Alykhan Tejani, Frank Portman, Alexandre Lung-Yut-Fong, Ben Chamberlain, Yuanpu Xie, Jonathan Hunt, Michael M. Bronstein, and Wenzhe Shi. 2021. RecSys 2021 Challenge Workshop: Fairness-aware engagement prediction at scale on Twitter’s Home Timeline. In RecSys ’21: Fifteenth ACM Conference on Recommender Systems, Amsterdam, The Netherlands, 27 September 2021 - 1 October 2021, Humberto Jesús Corona Pampín, Martha A. Larson, Martijn C. Willemsen, Joseph A. Konstan, Julian J. McAuley, Jean Garcia-Gathright, Bouke Huurnink, and Even Oldridge (Eds.). ACM, 819–824. https://doi.org/10.1145/3460231.3478515Google ScholarGoogle Scholar
  2. Luca Belli, Sofia Ira Ktena, Alykhan Tejani, Alexandre Lung-Yut-Fon, Frank Portman, Xiao Zhu, Yuanpu Xie, Akshay Gupta, Michael Bronstein, Amra Delić, Gabriele Sottocornola, Walter Anelli, Nazareno Andrade, Jessie Smith, and Wenzhe Shi. 2020. Privacy-Preserving Recommender Systems Challenge on Twitter’s Home Timeline. (2020). arxiv:2004.13715 [cs.SI]Google ScholarGoogle Scholar
  3. Luca Belli, Alykhan Tejani, Frank Portman, Alexandre Lung-Yut-Fong, Ben Chamberlain, Yuanpu Xie, Kristian Lum, Jonathan Hunt, Michael Bronstein, Vito Walter Anelli, Saikishore Kalloori, Bruce Ferwerda, and Wenzhe Shi. 2021. The 2021 RecSys Challenge Dataset: Fairness is not optional. arxiv:2109.08245 [cs.SI]Google ScholarGoogle Scholar
  4. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arxiv:1810.04805 [cs.CL]Google ScholarGoogle Scholar
  5. Peter Flach and Meelis Kull. 2015. Precision-recall-gain curves: PR analysis done right. In Advances in neural information processing systems. 838–846.Google ScholarGoogle Scholar
  6. Pere Gilabert and Santi Seguí. 2020. Gradient Boosting and Language Model Ensemble for Tweet Recommendation. In Proceedings of the Recommender Systems Challenge 2020. 24–28.Google ScholarGoogle Scholar
  7. Shuhei Goda, Naomichi Agata, and Yuya Matsumura. 2020. A Stacking Ensemble Model for Prediction of Multi-type Tweet Engagements. In Proceedings of the Recommender Systems Challenge 2020. 6–10.Google ScholarGoogle Scholar
  8. Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. (2017), 3146–3154. http://papers.nips.cc/paper/6907-lightgbm-a-highly-efficient-gradient-boosting-decision-tree.pdfGoogle ScholarGoogle Scholar
  9. Benedikt Schifferer, Gilberto Titericz, Chris Deotte, Christof Henkel, Kazuki Onodera, Jiwei Liu, Bojan Tunguz, Even Oldridge, Gabriel De Souza Pereira Moreira, and Ahmet Erdem. 2020. GPU Accelerated Feature Engineering and Training for Recommender Systems. In Proceedings of the Recommender Systems Challenge 2020. 16–23.Google ScholarGoogle Scholar
  10. Maksims Volkovs, Zhaoyue Cheng, Mathieu Ravaut, Hojin Yang, Kevin Shen, Jin Peng Zhou, Anson Wong, Saba Zuberi, Ivan Zhang, Nick Frosst, 2020. Predicting Twitter Engagement With Deep Language Models. In Proceedings of the Recommender Systems Challenge 2020. 38–43.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    RecSysChallenge '21: Proceedings of the Recommender Systems Challenge 2021
    October 2021
    43 pages
    ISBN:9781450386937
    DOI:10.1145/3487572

    Copyright © 2021 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 22 November 2021

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate11of15submissions,73%
  • Article Metrics

    • Downloads (Last 12 months)26
    • Downloads (Last 6 weeks)2

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format