skip to main content
column

A Study of the Overlap Among Document Representations

Published:02 August 2017Publication History
Skip Abstract Section

Abstract

Most previous investigations comparing the performance of different representations have used recall and precision as performance measures. However, there is evidence to show that these measures are insensitive to an important difference between representations. To explain, two representations may perform similarly on these measures, while retrieving very different sets of documents. Equivalence of representations should be decided on the basis of similarity in performance and similarity in the documents retrieved. This study compared the performance of four representations in the PsycAbs database. In addition, overlap between retrieved sets was also computed where overlap is the proportion of retrieved documents that are the same for pairs of document representations. Results indicate that for any two representations considered, performance values differed slightly while overlap scores were also low, thus supporting the evidence that recall and precision as performance measures mask differences between the sets of retrieved documents. Results are interpreted to propose an optimal ordering of the representations and to examine the contribution of each representation given this combination.

References

  1. Katzer, Jeffrey, et al. A Study of the Overlap Among Document Representations. Information Technology, 1982. i, pp. 201--273.Google ScholarGoogle Scholar
  2. Katzer, Jeffrey. A Study of the Impact of Representations in Information Retrieval Systems. Final report for Grant NSF-IST- 79-21468 to the National Science Foundation, July 1982.Google ScholarGoogle Scholar
  3. McGill, Michael J. et al. An Evaluation of Factors Affecting Document Ranking by Information Retrieval Systems. Final Report For Grant NSF-IST-78-10454 to the National Science Foundation, October 1979.Google ScholarGoogle Scholar

Index Terms

  1. A Study of the Overlap Among Document Representations
            Index terms have been assigned to the content through auto-classification.

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in

            Full Access

            • Published in

              cover image ACM SIGIR Forum
              ACM SIGIR Forum  Volume 51, Issue 2
              SIGIR Test-of-Time Awardees 1978-2001
              July 2017
              276 pages
              ISSN:0163-5840
              DOI:10.1145/3130348
              • Editors:
              • Donna Harman,
              • Diane Kelly
              Issue’s Table of Contents

              Copyright © 2017 Copyright is held by the owner/author(s)

              Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

              Publisher

              Association for Computing Machinery

              New York, NY, United States

              Publication History

              • Published: 2 August 2017

              Check for updates

              Qualifiers

              • column

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader