Abstract
Literature recommendation systems (LRS) assist readers in the discovery of relevant content from the overwhelming amount of literature available. Despite the widespread adoption of LRS, there is a lack of research on the user-perceived recommendation characteristics for fundamentally different approaches to content-based literature recommendation. To complement existing quantitative studies on literature recommendation, we present qualitative study results that report on users’ perceptions for two contrasting recommendation classes: (1) link-based recommendation represented by the Co-Citation Proximity (CPA) approach, and (2) text-based recommendation represented by Lucene’s MoreLikeThis (MLT) algorithm. The empirical data analyzed in our study with twenty users and a diverse set of 40 Wikipedia articles indicate a noticeable difference between text- and link-based recommendation generation approaches along several key dimensions. The text-based MLT method receives higher satisfaction ratings in terms of user-perceived similarity of recommended articles. In contrast, the CPA approach receives higher satisfaction scores in terms of diversity and serendipity of recommendations. We conclude that users of literature recommendation systems can benefit most from hybrid approaches that combine both link- and text-based approaches, where the user’s information needs and preferences should control the weighting for the approaches used. The optimal weighting of multiple approaches used in a hybrid recommendation system is highly dependent on a user’s shifting needs.
M. Ostendorff and C. Breitinger—Both authors contributed equally to this research.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
We judge the article quality using Wikipedia’s vital article policy [40].
- 2.
References
Apache Lucence - More Like This. https://lucene.apache.org/core/7_2_0/queries/org/apache/lucene/queries/mlt/MoreLikeThis.html. Accessed 16 June 2021
Bär, D., Zesch, T., Gurevych, I.: A reflective view on text similarity. In: Proceedings of the International Conference Recent Advances in Natural Language Processing 2011, pp. 515–520. Association for Computational Linguistics, Hissar (2011). https://www.aclweb.org/anthology/R11-1071
Beel, J., Breitinger, C., Langer, S., Lommatzsch, A., Gipp, B.: Towards reproducibility in recommender-systems research. User Model. User-Adap. Inter. 26(1), 69–101 (2016). https://doi.org/10.1007/s11257-016-9174-x
Beel, J., Gipp, B., Langer, S., Breitinger, C.: Research-paper recommender systems: a literature survey. Int. J. Digit. Libr. 17(4), 305–338 (2015). https://doi.org/10.1007/s00799-015-0156-0
Cremonesi, P., Garzotto, F., Negro, S., Papadopoulos, A.V., Turrin, R.: Looking for “Good’’ recommendations: a comparative evaluation of recommender systems. In: Campos, P., Graham, N., Jorge, J., Nunes, N., Palanque, P., Winckler, M. (eds.) INTERACT 2011. LNCS, vol. 6948, pp. 152–168. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-23765-2_11
De Gemmis, M., Lops, P., Semeraro, G., Musto, C.: An investigation on the serendipity problem in recommender systems. Inf. Process. Manag. 51(5), 695–717 (2015). https://doi.org/10.1016/j.ipm.2015.06.008
ElasticSearch - More like this query. https://www.elastic.co/guide/en/elasticsearch/reference/current/query-dsl-mlt-query.html. Accessed 16 June 2021
Färber, M., Sampath, A.: HybridCite: a hybrid model for context-aware citation recommendation. arXiv (2020). http://arxiv.org/abs/2002.06406
Ge, M., Delgado-Battenfeld, C., Jannach, D.: Beyond accuracy: evaluating recommender systems by coverage and serendipity. In: Proceedings of the Fourth ACM Conference on Recommender Systems, RecSys ’10, pp. 257–260. Association for Computing Machinery, New York (2010). https://doi.org/10.1145/1864708.1864761
Gipp, B., Beel, J.: Citation Proximity Analysis (CPA) – a new approach for identifying related work based on co-citation analysis. In: Proceedings of the 12th International Conference on Scientometrics and Informetrics (ISSI ’09), 2 July, pp. 571–575 (2009). https://ag-gipp.github.io/bib/preprints/gipp09a.pdf
Gravino, P., Monechi, B., Loreto, V.: Towards novelty-driven recommender systems. C R Phys. 20(4), 371–379 (2019). https://doi.org/10.1016/j.crhy.2019.05.014
Jannach, D., Zanker, M., Felfernig, A., Friedrich, G.: Recommender Systems - An Introduction, 1st edn. Cambridge University Press, London (2010)
Jinha, A.E.: Article 50 million: an estimate of the number of scholarly articles in existence. Learn. Publish. 23(3), 258–263 (2010). https://doi.org/10.1087/20100308
Joachims, T., Granka, L., Pan, B.: Accurately interpreting clickthrough data as implicit feedback. In: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (2005). https://doi.org/10.1145/1076034.1076063
Johnson, R., Watkinson, A., Mabe, M.: The STM report: an overview of scientific and scholarly publishing. Int. Assoc. Sci. Tech. Med. Publish. 5(October), 212 (2018)
Jones, K.S.: Index term weighting. Inf. Storage Retrieval 9(11), 619–633 (1973). https://doi.org/10.1016/0020-0271(73)90043-0
Kaminskas, M., Bridge, D.: Measuring surprise in recommender systems. In: RecSys REDD 2014: International Workshop on Recommender Systems Evaluation: Dimensions and Design, vol. 69, pp. 2–7 (2014)
Kaminskas, M., Bridge, D.: Diversity, serendipity, novelty, and coverage: a survey and empirical analysis of beyond-accuracy objectives in recommender systems. ACM Trans. Interact. Intell. Syst. 7(1), 1–42 (2017). https://doi.org/10.1145/2926720
Kanakia, A., Shen, Z., Eide, D., Wang, K.: A scalable hybrid research paper recommender system for microsoft academic. In: The World Wide Web Conference (WWW 2019), pp. 2893–2899. ACM Press, New York (2019). https://doi.org/10.1145/3308558.3313700
Knijnenburg, B.P., Willemsen, M.C., Gantner, Z., Soncu, H., Newell, C.: Explaining the user experience of recommender systems. User Model. User-Adap. Inter. 22(4–5), 441–504 (2012). https://doi.org/10.1007/s11257-011-9118-4
Knoth, P., Khadka, A.: Can we do better than co-citations? - Bringing citation proximity analysis from idea to practice in research article recommendation. In: 2nd Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries, Tokyo, Japan (2017)
Kunaver, M., Požrl, T.: Diversity in recommender systems - a survey. Knowl.-Based Syst. 123, 154–162 (2017). https://doi.org/10.1016/j.knosys.2017.02.009
Kunkel, J., Donkers, T., Michael, L., Barbu, C.M., Ziegler, J.: Let me explain: impact of personal and impersonal explanations on trust in recommender systems. In: Conference on Human Factors in Computing Systems - Proceedings (2019). https://doi.org/10.1145/3290605.3300717
Landis, J.R., Koch, G.G.: The measurement of observer agreement for categorical data. Biometrics 33(1), 159–174 (1977). http://www.jstor.org/stable/2529310
Lenhart, P., Herzog, D.: Combining content-based and collaborative filtering for personalized sports news recommendations. In: Proceedings of the 3rd Workshop on New Trends in Content-Based RecommenderSystems (CBRecSys ’16) at RecSys’16, vol. 1673, pp. 3–10 (2016)
Liu, S., Chen, C.: The effects of co-citation proximity on co-citation analysis. In: Proceedings of the 13th Conference of the International Society for Scientometrics and Informetrics (ISSI 2011) (2011)
Maksai, A., Garcin, F., Faltings, B.: Predicting online performance of news recommender systems through richer evaluation metrics. In: Proceedings of the 9th ACM Conference on Recommender Systems, pp. 179–186. ACM, New York (2015). https://doi.org/10.1145/2792838.2800184
Media Wiki Extension - CirrusSearch. https://www.mediawiki.org/wiki/Extension:CirrusSearch. Accessed 16 June 2021
Mendoza, M., Torres, N.: Evaluating content novelty in recommender systems. J. Intell. Inf. Syst. 54(2), 297–316 (2019). https://doi.org/10.1007/s10844-019-00548-x
Molloy, P., Beel, J., Aizawa, A.: Virtual Citation Proximity (VCP): learning a hypothetical in-text citation-proximity metric for uncited documents. In: Proceedings of the 8th International Workshop on Mining Scientific Publications, pp. 1–8 (2020). https://doi.org/10.31219/osf.io/t5aqf
Ostendorff, M., Ash, E., Ruas, T., Gipp, B., Moreno-Schneider, J., Rehm, G.: Evaluating document representations for content-based legal literature recommendations. In: Proceedings of the Eighteenth International Conference on Artificial Intelligence and Law, ICAIL ’21, pp. 109–118. Association for Computing Machinery, New York (2021). https://doi.org/10.1145/3462757.3466073
Pu, P., Chen, L., Hu, R.: A user-centric evaluation framework for recommender systems. In: Proceedings of the Fifth ACM Conference on Recommender Systems - RecSys ’11, p. 157. ACM Press, New York (2011). https://doi.org/10.1145/2043932.2043962
Salton, G., Wong, A., Yang, C.S.: Vector space model for automatic indexing. Information retrieval and language processing. Commun. ACM 18(11), 613–620 (1975)
Schwarzer, M., Breitinger, C., Schubotz, M., Meuschke, N., Gipp, B.: Citolytics: a link-based recommender system for wikipedia. In: Proceedings of the Eleventh ACM Conference on Recommender Systems - RecSys ’17, pp. 360–361. ACM, New York (2017). https://doi.org/10.1145/3109859.3109981
Schwarzer, M., Schubotz, M., Meuschke, N., Breitinger, C., Markl, V., Gipp, B.: Evaluating Link-based recommendations for wikipedia. In: Proceedings of the 16th ACM/IEEE-CS on Joint Conference on Digital Libraries - JCDL ’16, pp. 191–200. ACM Press, New York (2016). https://doi.org/10.1145/2910896.2910908
Sinha, R., Swearingen, K.: The role of transparency in recommender systems. In: CHI ’02 Extended Abstracts on Human Factors in Computing Systems - CHI ’02, p. 830. ACM Press, New York (2002)
Small, H.: Co-citation in the scientific literature: a new measure of the relationship between two documents. J. Am. Soc. Inf. Sci. 24(4), 265–269 (1973). https://doi.org/10.1002/asi.4630240406
Sparck Jones, K., Walker, S., Robertson, S.E.: A probabilistic model of information retrieval: development and comparative experiments, Part 2. Inf. Process. Manag. 36(6), 809–840 (2000). https://doi.org/10.1016/S0306-4573(00)00015-7
Wikipedia - Size of Wikipedia. https://en.wikipedia.org/wiki/Wikipedia:Size_of_Wikipedia. Accessed 16 June 2021
Wikipedia - Vital articles. https://en.wikipedia.org/wiki/Wikipedia:Vital_articles. Accessed 16 June 2021
Yu, C., Lakshmanan, L., Amer-Yahia, S.: It takes variety to make a world: diversification in recommender systems. In: Edbt, pp. 368–378 (2009). https://doi.org/10.1145/1516360.1516404
Zhang, Y., Chen, X.: Explainable recommendation: a survey and new perspectives. Now Publish. (2020). https://doi.org/10.1561/9781680836592
Zhao, Q., Harper, F.M., Adomavicius, G., Konstan, J.A.: Explicit or implicit feedback? Engagement or satisfaction? a field experiment on machine-learning-based recommender systems. In: Proceedings of the 33rd Annual ACM Symposium on Applied Computing, SAC ’18, pp. 1331–1340. Association for Computing Machinery, New York (2018). https://doi.org/10.1145/3167132.3167275
Acknowledgements
We thank the anonymous reviewers, all participants, and especially Olha Yarikova for her assistance in the completion of the interviews.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Ostendorff, M., Breitinger, C., Gipp, B. (2021). A Qualitative Evaluation of User Preference for Link-Based vs. Text-Based Recommendations of Wikipedia Articles. In: Ke, HR., Lee, C.S., Sugiyama, K. (eds) Towards Open and Trustworthy Digital Societies. ICADL 2021. Lecture Notes in Computer Science(), vol 13133. Springer, Cham. https://doi.org/10.1007/978-3-030-91669-5_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-91669-5_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-91668-8
Online ISBN: 978-3-030-91669-5
eBook Packages: Computer ScienceComputer Science (R0)