An empirical evaluation of the User Engagement Scale (UES) in online news environments

https://doi.org/10.1016/j.ipm.2015.03.003Get rights and content

Highlights

  • We examined the robustness of the User Engagement Scale (UES).

  • Three studies were conducted in Canada and the United Kingdom.

  • The UES sub-scales were reliable across three samples of online news browser.

  • A four-factor structure, rather than six, may be more appropriate.

  • The UES differentiated between online news sources and experimental conditions.

Abstract

Questionnaires are commonly used to measure attitudes toward systems and perceptions of search experiences. Whilst the face validity of such measures has been established through repeated use in information retrieval research, their reliability and wider validity are not typically examined; this threatens internal validity. The evaluation of self-report questionnaires is important not only for the internal validity of studies and, by extension, increased confidence in the results, but also for examining constructs of interest over time and across different domains and systems.

In this paper, we look at a specific questionnaire, the User Engagement Scale (UES), for its robustness as a measure. We describe three empirical studies conducted in the online news domain and investigate the reliability and validity of the UES. Our results demonstrate good reliability of the UES sub-scales; however, we argue that a four-factor structure may be more appropriate than the original six-factor structure proposed in earlier work. In addition, we found evidence to suggest that the UES can differentiate between systems (in this case, online news sources) and experimental conditions (i.e., the type of media used to present online content).

Introduction

In information interaction research, there is growing recognition for information systems to satisfy both utilitarian and hedonic needs, particularly in the area of exploratory search where users’ motivations go beyond merely completing an information task (White & Roth, 2009). This has created a call to action for researchers to design systems that are fun and engaging and that foster learning and discovery (White & Roth, 2009), but also to systematically consider the measurement of such experiences. Kelly (2009) underscores the need to capture affective responses during “particular IR [information retrieval] interactions and states” (p. 200), while O’Brien and Lebow (2013) advocate developing robust measures of experience. There is an increasing impetus to support searchers to not only retrieve and evaluate information, but to engage with it. Further, we need “good measures” that capture searchers’ affective experiences and allow a degree of standardization to facilitate comparisons between different search systems, user groups, etc. (Käki & Aula, 2008, p. 86).

Questionnaires are commonly used to measure attitudes toward systems and perceptions of search experiences. Kelly (2009) notes that such measures have good face validity, but are often developed “ad hoc” and their reliability and wider validity are not well established. This is problematic because, without reliable and valid measures, how can we be confident of our findings (Cairns, 2013)? Thus it is essential to evaluate the robustness of self-report measures to bolster internal validity.

This paper looks to a particular questionnaire, the User Engagement Scale (UES) to evaluate its potential as a “good measure” for adequately assessing human information experiences. The UES is a 31-item self-report questionnaire developed in the e-shopping domain (O’Brien & Toms, 2010a). It has been subject to some evaluation in search systems (Arguello et al., 2012, O’Brien and Toms, 2013), multimedia presentation software (O’Brien & Toms, 2010b), social networking applications (Banhawi & Mohamed Ali, 2011), and games (Wiebe, Lamb, Hardy, & Sharek, 2014) with mixed results regarding its generalizability in experimental settings, and the number of user engagement dimensions (UE) is captures. Given these mixed results, our motivating question was: “Is the UES a reliable and valid experiential scale to use in information interaction research, specifically online news browsing?” Here, we investigate its viability across three studies.

Section snippets

User engagement

User engagement (UE) encompasses users’ initial reactions to technologies (Sutcliffe, 2010), as well as sustained use of and re-engagement with information systems over time (Jacques, 1996, O’Brien and Toms, 2008). Engaging experiences involve system feedback and challenge, novelty, aesthetic and sensory appeal, interactivity, interest, choice, control, motivation, and positive affect (Jacques, 1996, Jacques et al., 1995; O’Brien, 2008; O’Brien and Toms, 2008, Webster and Ho, 1997). UE has

Current studies

Online news was the domain selected for these studies because it is interactive and multimodal (Chung, 2008), is widely used by a cross-section of the population for information seeking and browsing (Marshall, 2007), and satisfies both purposeful and serendipitous information needs. Online news, with its relevance to everyday life, ubiquity, and the interactive and multimodal nature of news delivery – is a fitting environment for studying UE and the utilitarian and hedonic aspects of

Methodology

The purpose of this study was to examine the plausibility and reliability of deploying the UES in the news domain to ensure that the UES retains its construct validity in this new domain. The design of this study was distinguished from the original UES work in several important ways. First, it examined the scale in the online news domain, rather than online shopping. This required the adaptation of the UES to be more generically suitable for engagement in a variety of domains. Also, it was

Study 2: UES and news source familiarity

In study 2, we re-examined the reliability of the UES and investigated its sensitivity by manipulating news source familiarity. There are many contextual factors that may influence UE with news, notably trust and credibility. Johnson and Kaye (2009), for example, explored politically oriented Internet users’ evaluations of political candidate websites, issues-oriented websites, blogs, email lists/bulletin boards and chat/instant sources. Users’ motivations to rely on blogs for non-mainstream

Study 3

Building on studies 1 and 2, we investigated the reliability and construct validity of the UES further, specifically its ability to detect differences between conditions. The overarching question guiding this research was whether the modality of interaction with news content would influence users’ level of engagement.

Previous research has demonstrated that video is more engaging for people interacting with digital media (Chapman et al., 1999), and that it may enhance users’ performance (Jacques

Implications for using the UES

We conducted three studies using the UES in the online news domain to examine its utility as a measure for discriminating online news experiences. In terms of reliability, all three studies showed support for the internal consistency of the six subscales prior to PCA (with the exception of NO and FI in Study 1, though this may be due to the small sample size). However, a four- (Study 3) or five-factor (Studies 1 and 2) solution emerged during PCA. AE, FA and PUs emerged as distinct factors in

Conclusion

In conclusion, we found evidence to support the reliability and construct validity of the UES across three studies conducted in the online news domain. While our PCA did not confirm the six-factor structure of O’Brien and Toms’ (2010a) original work, we did find support for the internal consistency of items with the subscales, and for PUs, AE and FA being stable factors. Our findings around the NO, FI and EN subscales is more perplexing: in Study 1 novelty emerged as a distinct factor, while in

Acknowledgments

We wish to thank the people who participated in the research studies. The first author (O’Brien) acknowledges the support of the Social Science and Humanities Council of Canada and the Networks of Centres of Excellence Graphics, Animation and New Media Project.

References (41)

  • A. Al-Maskari et al.

    The effect of user characteristics on search effectiveness in information retrieval

    Information Processing and Management

    (2011)
  • J. Arguello et al.

    Task complexity, vertical display and user interaction in aggregated search

  • F. Banhawi et al.

    Measuring user engagement attributes in social networking application

  • P. Cairns

    A critique of the UMUX usability questionnaire

    Interacting with Computers

    (2013)
  • P. Chapman et al.

    Engagement in multimedia training systems

  • D.S. Chung

    Interactive features of online newspapers: Identifying patterns and predicting use of engaged readers

    Journal of Computer-Mediated Communication

    (2008)
  • E.P. Cox

    The optimal number of response alternatives for a scale: A review

    Journal of Marketing Research

    (1980)
  • R.F. DeVellis

    Scale development: Theory and applications

    (2003)
  • S. Fulmer et al.

    A review of self-report and alternative approaches in the measurement of student motivation

    Educational Psychology Review

    (2009)
  • Hong, L., & Yang, S-H. (Eds.), (2013). Proceedings of the 1st workshop on user engagement optimization. 22nd ACM...
  • Hyder, J. A. (2010). Proposal of a website engagement scale and research model: Analysis of the influence of...
  • Jacques, R. D. (1996). The nature of engagement and its role in hypermedia evaluation and design. Doctoral Thesis....
  • R. Jacques et al.

    Engagement as a design concept for multimedia

    Canadian Journal of Education and Communication

    (1995)
  • T.J. Johnson et al.

    In blog we trust? Deciphering credibility of components of the internet among politically interested internet users

    Computers in Human Behavior

    (2009)
  • M. Käki et al.

    Controlling the complexity in comparing search user interfaces via user studies

    Information Processing and Management

    (2008)
  • D. Kelly

    Methods for evaluating interactive information retrieval systems with users

    Foundations and Trends in Information Retrieval

    (2009)
  • Kelly, D., Fu, X., & Shah, C. (2010). Effects of position and number of relevant documents retrieved on users’...
  • D. Kelly et al.

    Questionnaire mode effects in interactive information retrieval experiments

    Information Processing and Management

    (2008)
  • P. Kline

    A psychometrics primer

    (2000)
  • J. Lehmann et al.

    Models of user engagement

  • Cited by (55)

    • Does gamification engage users in online shopping?

      2021, Electronic Commerce Research and Applications
      Citation Excerpt :

      Nevertheless, scholars have rarely employed the UES in its entirety (O'Brien, 2018). Usually, some of its items have been removed, and the number of factors shaping the scale have been reduced (Banhawi and Ali, 2011; Wiebe et al., 2014; O’Brien and Cairns, 2015). This implies that evaluating UES generalizability becomes an arduous task.

    • Quantifying and Measuring Confirmation Bias in Information Retrieval Using Sensors

      2023, UbiComp/ISWC 2023 Adjunct - Adjunct Proceedings of the 2023 ACM International Joint Conference on Pervasive and Ubiquitous Computing and the 2023 ACM International Symposium on Wearable Computing
    View all citing articles on Scopus
    View full text