Elsevier

Journal of Informetrics

Volume 7, Issue 4, October 2013, Pages 897-906
Journal of Informetrics

Journal acceptance rates: A cross-disciplinary analysis of variability and relationships with journal measures

https://doi.org/10.1016/j.joi.2013.08.007Get rights and content

Highlights

  • Acceptance rates vary statistically significantly by discipline of journal, country affiliation of the editor, and number of reviewers per article.

  • There is a negative relationship between acceptance rates and citation-based indicators.

  • There is a positive relationship between acceptance rates and journal age.

  • Open access journals have significantly higher mean acceptance rates than non-open access journals.

Abstract

There are many indicators of journal quality and prestige. Although acceptance rates are discussed anecdotally, there has been little systematic exploration of the relationship between acceptance rates and other journal measures. This study examines the variability of acceptance rates for a set of 5094 journals in five disciplines and the relationship between acceptance rates and JCR measures for 1301 journals. The results show statistically significant differences in acceptance rates by discipline, country affiliation of the editor, and number of reviewers per article. Negative correlations are found between acceptance rates and citation-based indicators. Positive correlations are found with journal age. These relationships are most pronounced in the most selective journals and vary by discipline. Open access journals were found to have statistically significantly higher acceptance rates than non-open access journals. Implications in light of changes in the scholarly communication system are discussed.

Introduction

The scholarly publication system operates on the basis of exchange. As in any market, there are suppliers (authors) and buyers2 (journals) of goods (papers). In this exchange, authors are seeking journals with the highest impact in order to increase their stock of symbolic capital (Bourdieu, 1984), while journals attempt to capture papers that will increase their prestige. As a general rule, the best authors want to publish in the best journals, and the best journals want the best authors (those with the potentially best papers) to publish with them. In a perfect (i.e., optimally efficient) market, the best papers would gravitate to the best journals (Oster, 1980). But in this as so many other markets both suppliers and buyers lack perfect information. Absent perfect information, the various actors involved rely upon a range of indicators (bibliometric, sociometric, and demographic) and tacit knowledge to guide decision-making. One such indicator is acceptance rate, the proportion of papers submitted to a journal that are subsequently accepted and published.

Space at the upper end of the market is highly sought after and in limited supply. Competition and ambition often drive scholars to submit papers to journals beyond their reach, creating a cascade of rejected papers that puts added pressure on reviewers and editors (Craig, 2010, Cronin and McKenzie, 1992, Kravitz and Baker, 2011). Economic models have been proposed to analyze, inter alia, research spillover effects, duality in scientific discovery and congestion in information processing (Besancenot, Huynh, & Vranceanu, 2009, p. 1). Such models highlight the “informational frictions” that occur when papers are being matched with journals (Besancenot et al., 2009, p. 2).

Peer review is the established mechanism for allocating space to papers within a journal. Experts (editors, editorial board members and external reviewers) assess the quality of submitted papers and evaluate their suitability for publication. It is assumed that editors and reviewers are unbiased in their assessments and that the governing norm of impartiality is not violated, at least not egregiously (Lee et al., 2013, Sugimoto and Cronin, 2013). In reality, it is not quite so straightforward, as variations in consensus as to what constitutes quality, broadly conceived, within and across fields, can have an effect on acceptance rates (Hargens, 1988, Kravitz et al., 2010).

Variation in journal acceptance rates is an understudied area, not the least because of the difficulty in obtaining reliable data. One of the most comprehensive (and earliest) studies to date examined the rejection rates of 83 journals across a broad spectrum of disciplinary areas and found that humanities and social science journals have the highest and the biological sciences the lowest rates of rejection (Zuckerman & Merton, 1971, p. 77): “the more humanistically oriented the journal, the higher the rate of rejecting manuscripts for publication; the more experimentally oriented, with an emphasis on rigor of observation and analysis, the lower the rate of rejection.” Subsequent monodisciplinary studies have confirmed these findings (e.g., Cherkashin et al., 2009, Rotton et al., 1993, Schultz, 2010, Seaton, 1975, Vlachy, 1981). One explanation for this is the degree to which a dominant paradigm exists in any given discipline, providing a consensus as to what constitutes valid research (Kuhn, 1970).

It has been noted that there exists little guidance for calculating acceptance rates (Moore & Perry, 2012). At face value, the calculation may seem simple enough—the number of papers accepted over the total number of papers submitted. However, this is complicated by the unreliability of self-report data, inconsistent definitions of a resubmission, the inclusion/exclusion of invited papers or special issues in the calculations, the timeframe used, and the inclusion/exclusion of book reviews, among other considerations (Moore & Perry, 2012). Additionally, many studies rely on individual surveys of editors/publishers, rather than using a standard source for evaluation. Cabell's Directories of Publishing Opportunities (Cabell's henceforth) is one such source, but has been used only rarely in empirical studies (e.g., Haensly, Hodges, & Davenport, 2008).

Acceptance rates ostensibly testify to the relative competitiveness of a journal and have been used as a quality indicator. Statistically significant negative correlations between acceptance rates and other proxies of quality (i.e., citation rates, Journal Impact Factor [JIF]) have been demonstrated (Buffardi and Nichols, 1981, Haensly et al., 2008, Lee et al., 2002). Rotton et al. (1993) found that rejection rates were good predictors of citations, while Haensly et al. (2008) found acceptance rates to be statistically significantly correlated with both citations and survey-based rankings of journals. However, and with few exceptions, these have relied on small scale and monodisciplinary datasets and are somewhat dated.

More comprehensive studies are necessary to elucidate the relationship between acceptance rates and other indicators. The results of such studies can be used to assess the utility of journal acceptance rates and the degree to which these can be considered appropriate proxies for quality. To this end, the present paper provides the largest study of acceptance rates to date and investigates the following research questions and associated hypotheses:

  • 1.

    What is the degree of variability in acceptance rates?

    H1. Statistically significant differences in acceptance rates will be observable by discipline.

    H2. Statistically significant differences in acceptance rates will be observable by country of editor's location.

    H3. Statistically significant differences in acceptance rates will be observable by number of external reviewers.

    H4. Statistically significant differences in acceptances rates will be observable between JCR and non-JCR journals.

  • 2.

    What is the relationship between acceptance rates and JCR measures?

    H5. Statistically significant negative correlations will be observable between acceptance rates and citation-based JCR measures.

    H6. Statistically significant positive correlations will be observable between acceptance rates and size-based JCR measures (e.g., number of publications).

  • 3.

    What is the relationship between acceptance rates and age of journal?

    H7. Statistically significant negative relationships will be observable between acceptance rates and age of journal.

  • 4.

    What is the relationship between acceptance rates and the open access status of journals?

    H8. Open access journals will have statistically significantly higher acceptance rates.

The results of this analysis will inform scientometricians, policy makers, and scholars who employ these metrics to make decisions about where to publish and the relative quality of venues.

Section snippets

Data

We used four main sources of data: Cabell's, Thomson Reuters’ Journal Citation Reports (JCR) for both Science and Social Sciences, Ulrich's Periodicals Directory (Ulrich's), and the Directory of Open Access Journals.

Cabell's provides general descriptive information about journals (Cabell's Directories, 2013). It indexes journals in eleven specialties organized into five disciplines (Table 1); a journal can be assigned to multiple specialties. New journals can be recommended to the directory by

Results

The results are split into three sections. The first section describes and analyzes all journals indexed by Cabell's, using the fields available from this source. The second section describes and analyzes only those journals found in both Cabell's and the JCR and also incorporates journal age information from Ulrich's. The final section analyzes the difference in acceptance rates between open access and non-open access journals.

Discussion and conclusion

The first research question related to variability. Examining more than 5000 journals, statistically significant differences in acceptance rates were found by discipline, country affiliation of the editor, and number of reviewers. This variability suggests that acceptance rates, like many other scientometric indicators, are highly contextualized and should be used accordingly. Health journals had the highest acceptance rates—this may be a result of the particular specialties (i.e., Nursing and

References (38)

  • I. Cherkashin et al.

    The inside scoop: Acceptance and rejection at the Journal of International Economics

    Journal of International Economics

    (2009)
  • É. Archambault et al.

    Benchmarking scientific output in the social sciences and humanities: The limits of existing databases

    Scientometrics

    (2006)
  • P. Basken

    Researchers and scientific groups make new push against impact factors

    The Chronicle of Higher Education

    (2013)
  • D. Besancenot et al.

    Desk rejection in an academic publication market model with matching frictions. DR 09008

    (2009)
  • L. Bornmann et al.

    A content analysis of referees’ comments: How do comments on manuscripts rejected by a high-impact journal and later published in either a low- or high-impact journal differ?

    Scientometrics

    (2010)
  • P. Bourdieu

    Distinction: A social critique of the judgement of taste

    (1984)
  • L.C. Buffardi et al.

    Citation impact, acceptance rate, and APA journals

    American Psychologist

    (1981)
  • Cabell's Directories

    Directories

    (2013)
  • V. Calcagno et al.

    Flows of research manuscripts among scientific journals reveal hidden submission patterns

    Science

    (2012)
  • J. Cohen

    Statistical power analysis for the behavioral sciences

    (1988)
  • J.B. Craig

    Desk rejection: How to avoid being hit by a returning boomerang

    Family Business Review

    (2010)
  • B. Cronin

    Editorial. Do me a favor

    Journal of the American Society for Information Science and Technology

    (2012)
  • B. Cronin et al.

    The trajectory of rejection

    Journal of Documentation

    (1992)
  • A.R. Dennis et al.

    Research standards for promotion and tenure in information systems

    MIS Quarterly

    (2006)
  • Eigenfactor.org.

    Cost effectiveness for open access journals

    (2013)
  • P.J. Haensly et al.

    Acceptance rates and journal quality: An analysis of journals in economics and finance

    Journal of Business & Finance Librarianship

    (2008)
  • L.L. Hargens

    Scholarly consensus and journal rejection rates

    American Sociological Review

    (1988)
  • C. Harris

    Ranking the management journals

    Journal of Scholarly Publishing

    (2008)
  • Cited by (73)

    • Middle East Authors' Contribution to the Journal of Arthroplasty's Publications in the Past 20 years (2000–2020)

      2022, Arthroplasty Today
      Citation Excerpt :

      The need for scientific publications and research productivity for career advancement and achieving promotion is mandatory for orthopedic researchers, with no exception for arthroplasty surgeons [7,8]. Studies are preferably published in high-ranking peer-reviewed journals [9,10]; however, most of these journals, including the Journal of Arthroplasty (JOA), have low acceptance rates of approximately 15%–40% [11-13]. One way to measure scientific productivity is performing bibliometric studies, which are popularized in all scientific fields in general and the medical field in particular [14-16].

    • The emerging identity and reputation of SEPR

      2024, Socio-Ecological Practice Research
    View all citing articles on Scopus
    View full text