Skip to main content
Log in

How User Engagement Metrics Ameliorate the Web QoE?

  • Published:
Wireless Personal Communications Aims and scope Submit manuscript

Abstract

User engagement in general is an interpretation of an individual’s reaction to an offer (a service, a product, an application). We can determine a person’s degree of commitment directly by interacting with him or by observing his behavior. To measure the user engagement some actions can be interpreted such as clicks on links, comments and download of documents. In an other hand, we can measure the user satisfaction towards a web page using web QoE metrics such as Time to First Byte (TTFB), onLoad (Load Time) and Above the Fold (ATF). These metrics are commenly used to predict web QoE using Machine Learning (ML) algorithms. By comparing the two groups of metrics (user engagement and web QoE) we conclude that user engagaement metrics are closer to user desires, that’s why in our work we focus on user engagement to predict user’s satisfaction by predicting the Mean Opinion Score (MOS). Existing works dont use user engagement metrics in datasets to predict MOS, that’s why in this paper, we propose to focus on user engagement to predict our MOS. To do so, we use the help of an existing dataset whose parameters are web QoE metrics. At first, we select metrics that have a direct relation with user enagement. Then, we refine their coefficients to obtain the best combination that gives a MOS very close to the real one (expressed directly by users). After that, we add a new column to the existing dataset that containes the new obtained engagement parameter. Finally, we apply different ML algorithms on our new dataset to predict the MOS and we conclude that decision tree is the best in our case.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Kemp, S. Digital in 2018. Hootsuite Media Inc. https://hootsuite.com/fr/pages/digital-in-2018.

  2. Butkiewicz, M., & Madhyastha, H. V., et al. (2014). Characterizing web page complexity and its impact. In IEEE/ACM Transactions on Networking (pp. 943–956).

  3. Belshe, M, et al. (2015). Hypertext transfer protocol version 2 (http/2). In Internet Engineering Task Force.

  4. Wang, X. S., & Balasubramanian, A., et al. (2014). How speedy is spdy? In 11th USENIX Symposium on Networked Systems Design and Implementation.

  5. Wang, X. S., & Krishnamurthy, A., et al. (2016). Speeding up web page loads with shandian. In USENIX NSDI.

  6. Bocchi, E., & De Cicco, L., et al. (2016b). Measuring the quality of experience of web users. In ACM SIGCOMM CCR.

  7. Gao, Q., & Dey, P., et al. (2017a). Perceived performance of top retail webpages in the wild: Insights from large-scale crowdsourcing of above-the-fold qoe. In ACM Internet-QoE Workshop.

  8. Kelton, C., & Ryoo, J., et al. (2017). Improving user perceived page load time using gaze. In USENIX NSDI.

  9. Varvello, M., & Blackburn, J., et al. (2016). Eyeorg: A platform for crowdsourcing web quality of experience measurements. In ACM CoNEXT.

  10. Brutlag, J., & Abrams, Z., et al. (2011). Above the fold time: Measuring web page performance visually. In Web Performance and Operations.

  11. Quesenbery,W., & Whitney Interactive Design. (2003). Dimensions of usability: Defining the conversation, driving the process. In UPA (pp. 387–399).

  12. Bocchi, E., De Luca, C., & Rossi, D. (2016). Measuring the quality of experience of web users. In ACM SIGCOMM Computer Communication Review (pp. 8–13).

  13. Varvello, M., Blackburn, J., Naylor, D., & Papagiannaki, K. (2016). Eyeorg: A platform for crowdsourcing web quality of experience measurements. In ACM SIGCOMM Computer Communication Review (pp. 399–412).

  14. Barakovic, S., & Skorin-Kapov, L. (2017). Survey of research on quality of experience modelling for web browsing. Quality and User Experience. https://doi.org/10.1007/s41233-017-0009-2.

  15. Varela, M., Skorin-Kapov, L., Maki, T., & Hossfeld, T. (2015). Qoe in the web: A dance of design and performance. In IEEE QoMEX (pp. 1–7).

  16. Hobfeld, T., et al. (2011). The memory effect and its implications on web qoe modeling. In IEEE ITC (pp. 103–110).

  17. Egger, S., Reichl, P., Hoßfeld, T., & Schatz, R. (2012). "Time is bandwidth”? narrowing the gap between subjective time perception and quality of experience. In IEEE ICC (pp. 1325–1330).

  18. Strohmeier, D., Mikkola, M., & Alexander, R. (2013). The importance of task completion times for modeling web-qoe of consecutive web page requests. In IEEE QoMEX (pp. 38–39).

  19. Andreas, S., et al. (2015). Quantifying the impact of network bandwidth fluctuations and outages on web qoe. In IEEE QoMEX (pp. 1–6).

  20. Bocchi, E., Cicco, L. D., & Dario, R. (2016a). Measuring the quality of experience of web users. In Internet QoE Workshop (pp. 37–42).

  21. Alemnew, S. A., et al. (2019). Measuring web latency and rendering performance: Method, tools, and longitudinal dataset. IEEE Trasactions on Network and Service Management, 16, 535–549.

    Article  Google Scholar 

  22. Song, E., Pan, T., Qiang, F., et al. (2020). Threshold-oblivious on-line web qoe assessment using neural network-based regression model. IET Communications, 14, 2018–2026.

    Article  Google Scholar 

  23. Jahromi, H. Z., Delaney, T. D., & Hines, A. (2020). Beyond first impressions: Estimating quality of experience for interactive web applications. IEEE Access, 8, 47741–47755.

    Article  Google Scholar 

  24. Attfield, S., Kazai, G., Lalmas. M., & Piwowarski, B. (2011). Towards a science of user engagement (position paper). In WSDM Workshop on UMWA.

  25. Janett, L., et al. (2012). Models of user engagement. User modeling, adaptation, and personalization. UMAP 7379.

  26. Lagun, D., & Mounia, L. (2016). Understanding and measuring user engagement and attention in online news reading. In WSDM’16: Proceedings of the Ninth ACM International Conference on Web Search and Data Mining (pp. 113–122).

  27. Moldovan, C., & Florian, M. (2016). Bridging the gap between qoe and user engagement in http video streaming. In 28th International Teletraffic Congress: The First International Conference in Networking Science and Practice.

  28. Attfield, S., Kazai, G., Lalmas, M., Piwowarski, B. Towards a science of user engagement (position paper). In WSDM Workshop on User Modelling for Web Applications (2011).

  29. Grover, P., & Kar, A. K. (2020). User engagement for mobile payment service providers: Introducing the social media engagement model. Journal of Retailing and Consumer Services, 53. https://doi.org/10.1016/j.jretconser.2018.12.002

  30. Bonilla, M. R., Pereaa, E., del Olmo, J. L., & Corronsb, A. (2020). Insights into user engagement on social media. case study of a higher education institution. Journal of Marketing for Higher Education, 30, 145–160.

    Article  Google Scholar 

  31. Aung, W. T., Myanmar, Y., & Khin. H. (2009). Random forest classifier for multicategory classification of web pages. In IEEE Asia-Pacific Services Computing Conference (APSCC) (pp. 372–376).

  32. Pal, M., & Mather, P. (2002). A comparison of decision tree and back propagation neural network classifiers for land use classification. In IEEE International Geoscience and Remote Sensing Symposium c’IGARSS (pp. 503–505).

  33. Islam, M. J., Wu, Q. M. l, Ahmadi, M., & Sid-Ahmed, M. A. (2007). Investigating the performance of naive-bayes classifiers and k-nearest neighbor classifiers. In Convergence Information Technology, International Conference (pp. 1541–1546).

  34. Fiedler, M., Hossfeld, T., & Tran-Gia, P. (2010). A generic quantitative relationship between quality of experience and quality of service. In IEEE Network: The Magazine of Global Internetworking (pp. 36–41).

  35. Khan, A., Sun, L., Jammeh, E., & Ifeachor, E. (2010). Quality of experience driven adaptation scheme for video applications over wireless networks. In IET Communications (pp. 1337–1347).

  36. Samet, N., BenLetaifa, A., Hamdi, M., & Tabbane, S. T. (2016). Real-time user experience evaluation for cloud-based mobile video. In International Conference on Advanced Information Networking and Applications Workshops (WAINA) (pp. 204–208).

  37. Shahid, M., Rossholm, A., & Lövström, B. (2013). A no-reference machine learning based video quality predictor. In International Workshop on Quality of Multimedia Experience (QoMEX) (pp. 176–181).

  38. Le Callet, P., Viard-Gaudin, C., & Barba, D. (2006). A convolutional neural network approach for objective video quality assessment. In IEEE Transactions on Neural Networks (pp. 1316–1327).

  39. Menkovski, V., Exarchakos, G., & Liotta, A. (2010). Machine learning approach for quality of experience aware networks, In International Conference on Intelligent Networking and Collaborative Systems (pp. 461–466).

  40. Hameed, A., Dai, R., & Benjamin, B. (2016). A decision-tree-based perceptual video quality prediction model and its application in fec for wireless multimedia communications. In IEEE Transactions on Multimedia (pp. 764–774).

  41. Machado, V. A., & Silva, C. S., et al. (2011). A new proposal to provide estimation of qos and qoe over wimax networks. In IEEE Third Latin-American Conference on Communications.

  42. Mason, L., Baxter, J., Bartlett, P., & Frean, P. (1999). Boosting algorithms as gradient descent. In: International Conference on Neural Information Processing Systems (pp. 512–518).

  43. Kang, Y., Chen, H., & Lei, X. (2013). An artificial-neural-network-based qoe estimation model for video streaming over wireless networks. In IEEE/CIC International Conference on Communications in China (ICCCC): QRS: QoS, Reliability and Security (pp. 764–774).

  44. Wassermann, S., Wehner, N., & Casas, P. (2019). Machine learning models for youtube qoe and user engagement prediction in smartphones. ACM SIGMETRICS Performance Evaluation, 46, 155–158

  45. Mendez, A. E., Cartwright, M., & Juan, P. B. (2019). Machine-crowd-expert model for increasing user engagement and annotation quality. In CHI EA ’19 (pp. 1–6).

  46. Aluri, A., Price, B. S., & Nancy, H. M. (2019). Using machine learning to cocreate value through dynamic customer engagement in a brand loyalty program. Journal of Hospitality and Tourism Research, 43, 78–100.

    Article  Google Scholar 

  47. Chen, Y., Xiong, J., et al. (2018). A novel online incremental and decremental learning algorithm based on variable support vector machine. In Cluster Computing.

  48. Gao, Q., Dey, P., & Ahammad, P. (2017b). Perceived performance of top retail webpages in the wild: Insights from large-scale crowdsourcing of above-the-fold qoe. In Proceedings of the 2017 SIGCOMM Internet-QoE Workshop.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nawres Abdelwahed.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Abdelwahed, N., Ben Letaifa, A. & El Asmi, S. How User Engagement Metrics Ameliorate the Web QoE?. Wireless Pers Commun 117, 2383–2402 (2021). https://doi.org/10.1007/s11277-020-07980-1

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11277-020-07980-1

Keywords

Navigation