Measuring user perceptions of Web site reputation

https://doi.org/10.1016/j.ipm.2003.08.007Get rights and content

Abstract

In this study, we compare a search tool, TOPIC, with three other widely used tools that retrieve information from the Web: AltaVista, Google, and Lycos. These tools use different techniques for outputting and ranking Web sites: external link structure (TOPIC and Google) and semantic content analysis (AltaVista and Lycos). TOPIC purports to output, and highly rank within its hit list, reputable Web sites for searched topics. In this study, 80 participants reviewed the output (i.e., highly ranked sites) from each tool and assessed the quality of retrieved sites. The 4800 individual assessments of 240 sites that represent 12 topics indicated that Google tends to identify and highly rank significantly more reputable Web sites than TOPIC, which, in turn, outputs more than AltaVista and Lycos, but this was not consistent from topic to topic. Metrics derived from reputation research were used in the assessment and a factor analysis was employed to identify a key factor, which we call ‘repute’. The results of this research include insight into the factors that Web users consider in formulating perceptions of Web site reputation, and insight into which search tools are outputting reputable sites for Web users. Our findings, we believe, have implications for Web users and suggest the need for future research to assess the relationship between Web page characteristics and their perceived reputation.

Introduction

Multiple search engines retrieve and rank Web sites based on their semantic content and/or on their positioning within the link structure of the Web. These engines output from a dozen to a million or more sites on almost any topic, but often leave the user to sort and sift through the results looking for quality information. The search engines do a reasonable job of ranking Web sites according to their semantic content, but they do not incorporate into search algorithms indicators of quality. Assessing the quality of a Web site’s content is left to the individual, who often uses criteria other than topicality in his or her assessment (Rieh & Belkin, 1998).

One criterion used in evaluating Web sites is reputation. In general, reputation is an important personal asset in which to invest and protect (Weigelt & Camerer, 1988) and a strategic asset of public and private organizations (Fombrun, 2001). It can be considered an expectation of quality (Shapiro, 1982). In essence, one perceives through a variety of factors that a person or object has a good reputation and it raises expectations about the interactions one might have with that person or object. This situation is similar to interactions on the Web. For example, one has expectations about interactions with Amazon.com and those same expectations may not exist for a host of other online booksellers. On exposure to a Web site, however, a user may not know about the reputation or have perceptions about the reputation of that site and may mentally ask a number of questions: Is this “for real”? Can I trust it? Is it accurate? And so on. Users sometimes make split-second decisions about whether to stay and read, or move on quickly to other sites. This decision is based on how the individual perceives a whole series of visual and content cues emanating from the site. From this surface inspection, individuals derive an impression––they assess the quality of that site and thus, infer its reputation.

The volume of Web sites and the predicted growth of Web content preclude the ability to provide human indexing and ranking of Web sites for all given topics. Search engines must be able to automatically sift through this content and highly rank sites that are perceived as reputable to users, and not merely identify pages that are on a topic. At present, identifying good pages is left to users, who have to sift through results lists and the related Web sites, and make quick judgements about a page. In this research, we tested the ability of current search engines to output ‘reputable’ Web sites.

Unlike typical search tools, TOPIC (http://www.cs.toronto.edu/db/topic/search.html), developed by Rafiei and Mendelzon (2000), identifies the topics (or subjects) for which a Web site has a good reputation. It does so by mining both the link structure and the textual content of that site to indicate how well connected the site is to other well-connected sites on the Web. Thus, upon searching the Web for sites on a topic, TOPIC purports to output and highly rank sites that are highly perceived by the Web community for that topic.

To do this research, which is a follow-up to a pilot study (Keast, Toms, & Cherry, 2001), we develop a metric for testing the reputation of Web sites and compare the output from TOPIC with that of three other search engines (Google, AltaVista and Lycos) to ascertain if certain tools more often output and highly rank ‘reputable’ Web sites for given topics. Google (http://www.google.com) and TOPIC primarily rely on link structure analysis whereas AltaVista (http://www.altavista.com) and Lycos (http://www.lycos.com) depend on semantic content analysis when outputting and ranking Web sites for a search on a specific topic. In our pilot study, we included search engines and directories with human-selected and indexed content (Yahoo and Britannica.com). This third type was dropped in this larger study, as it performed no better than the automated tools in the pilot, and thus was deemed less essential to the intent of our study––the abilities of automatic indexing and ranking tools to output quality Web sites. Automated tools must be able to address the task of indexing and sifting through a large number of Web sites to allow users to access information they consider to be salient and of good quality. Thus, the work reported here assesses the ability of two types of indexing and ranking tools, represented by four search engines, to output Web sites that are perceived by Web users to be highly reputable on a specific topic.

Section snippets

What is reputation?

Reputation, according to the Oxford English Dictionary, is “the relative estimation or esteem in which a person or thing is held” (Simpson & Weiner, 1989, vol. XIII, p. 678). It is the external perception that humans have of another person or object, but it is not necessarily a true indication of that person’s or object’s qualities (Weigelt & Camerer, 1988). That said, a reputation might be positive or negative or simply neutral. It is used by people in their interactions with another person or

Overview

From the pilot study (Keast et al., 2001), we knew that assessing individual Web pages was a mentally taxing experience. Examining and evaluating a set of 20–30 sites in one sitting was about the maximum effort that a single person reasonably could expend on the task. In addition, we predicted that we would likely have attrition if participants were asked to return multiple times. In the design of our main study, we compromised by using a mixed between- and within-subjects design.

We selected 12

Overview

Eighty participants rated the 240 Web pages representing 12 topics, on average, 3.7, 3.9 and 3.8 for authoritativeness, trustworthiness and aboutness, respectively (using a five-point scale). They indicated that they would recommend 52%, and would return to 53% of the sites for more information on the topic. Overall, they placed about 31% of the sites in the “top five” sites for information on the topic.

Most of the sites were new to participants before the start of the study. Only 10% of the

Discussion and analysis

In this study, ordinary people assessed the quality of the output from four search tools, two of which were based on link structure analysis, and two on semantic content analysis. One of the link-based tools, TOPIC, purports to output reputable Web sites. The Web sites output by the four search engines were assessed using six metrics that were derived from quality assessments in other domains and were also considered to be indicators of reputation. In general, search tools based on link

Future work

In this study, people were asked to judge the quality of each Web site they were asked to examine. While the intention was to assess content, we do not know the extent to which the “window dressing” contributed to the evaluation. Weigelt and Camerer (1988) argue that game players who do not know a player’s true type will perceive a reputation from cues that the player emits. In our case, did some of those cues come from design elements, and not content? Design quality varies substantially from

Conclusions

We assessed the ability of four search tools to output reputable Web sites and know of no other work in which this particular attribute has been assessed. In general, output was perceived at just above neutral. Sites retrieved from Google were perceived to be of higher quality than the other three tools. TOPIC, which purports to output reputable Web sites, was a close second. In general, the link-based tools performed better than the semantic-based tools. But at the specific subject-matter

Acknowledgements

We thank the Faculty of Information Studies for the use of its facilities; the 80 participants; Alberto Mendelzon for providing funding for this project; and Joan Cherry and Greg Keast who worked on the pilot study. We also thank the anonymous reviewers for their very helpful comments and suggestions.

References (55)

  • Argus Associates. (2001). The Argus clearinghouse. Available:...
  • Bailey, B. P., Gurak, L. J., & Konstan, J. A. (2001). An examination of trust production in computer-mediated exchange....
  • Brin, S., & Page, L. (1998). The anatomy of a large-scale hypertextual Web search engine. In 7th international World...
  • M. Chen et al.

    Computing and using reputations for Internet ratings

  • W. Chu et al.

    Signalling quality by selling through a reputable retailer: an example of renting the reputation of another agent

    Marketing Science

    (1994)
  • C. Dellarocas

    Analyzing the economic efficiency of ebay-like online reputation reporting mechanisms

  • A. Dillon et al.

    It’s the journey and the destination: shape and the emergent property of genre in digital documents

    New Review of Multimedia and Hypermedia

    (1997)
  • V.J. Doby et al.

    Organizational stress as threat to reputation: effects on anxiety at work and at home

    The Academy of Management Journal

    (1995)
  • eBay. (2002). eBay––the world’s online marketplace. Available:...
  • Epinions. (2001). Epinions.com––welcome. Available:...
  • B.J. Fogg et al.

    What makes Web sites credible? A report on a large quantitative study

  • B.J. Fogg et al.

    The elements of computer credibility

  • C. Fombrun et al.

    What’s in a name? Reputation building and corporate strategy

    The Academy of Management Journal

    (1990)
  • Fombrun, C. J. (2001). Reputations: measurable, valuable, and manageable. American Banker Online, May 25, 2001....
  • C.W.L. Hill

    Cooperation, opportunism, and the invisible hand: implications for transaction cost theory

    The Academy of Management Review

    (1990)
  • L.T. Hosmer

    Trust: the connecting link between organizational theory and philosophical ethics

    The Academy of Management Review

    (1995)
  • C.I. Hovland et al.

    Communication and persuasion: Psychological studies of opinion change

    (1953)
  • Cited by (48)

    • Do reviewers’ words affect predicting their helpfulness ratings? Locating helpful reviewers by linguistics styles

      2019, Information and Management
      Citation Excerpt :

      Finally, the formation of a WoT requires members to identify others they trust in addition to providing product ratings, which might limit the applicability of the model. The ways in which a reviewer’s contribution can be earned include quantifiable knowledge of a domain, quality content, or social networks [16]. In response to concerns about using a social network approach, such as applying the WoT in estimating a reviewer’s helpfulness rating, this study turns to the factor of domain knowledge and quality content.

    • To whom should i listen? Finding reputable reviewers in opinion-sharing communities

      2012, Decision Support Systems
      Citation Excerpt :

      Finally, we conclude with a summary and discussion of some future research directions in Section 5. Reputation is defined as “the relative estimation or esteem in which a person or thing is held” [41] or “what is generally said or believed about a person's or thing's character or standing” [21]. In e-commerce environments, reputation refers to the shadow of the future, which supports each transaction made by buyers online.

    • Reputation rating modeling for open environment lack of communication by using online social cognition

      2012, Journal of Network and Computer Applications
      Citation Excerpt :

      There are arguments (Ennew et al., 2005; Gutowska and Sloane, 2010; Robins and Holmes, 2008; Yoo and Donthu, 2001) that conclusions can be drawn from ease of use, esthetics design, influence in areas, etc. Toms and Taves (2004) thought Web site trust and reputation evaluation could be achieved by searching engine and be used as the base of decision making (Artz and Gil, 2007). Therefore we adopted link speed, site ranking, Web click-through rate (CTR), amount of information in the search engines and Web pages' esthetics to collect information and build models to carry out the final rating evaluation.

    View all citing articles on Scopus
    View full text