skip to main content
10.1145/3366424.3383417acmconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article

Expanding the Scope of Reproducibility Research Through Data Analysis Replications

Published: 20 April 2020 Publication History

Abstract

In recent years, researchers in several scientific disciplines have become concerned with published studies replicating less often than expected. A positive side effect of this concern is an increased appreciation for replicating other researchers’ work as a vital part of the scientific process. To date, many such efforts have come from the experimental sciences, where replication entails running new experiments, generating new data, and analyzing it. In this article, we emphasize not experimental replication but data analysis replication. We do so for three reasons. First, experimental replication excludes entire classes of publications that do not run experiments or even collect original data (for example, papers that make use of economic data, census data, municipal data, and the like). Second, experimental replication may in some cases be a needlessly high bar: there is great value in replicating the data analyses of published experimental work. As analytical replications require a lower investment of time and money than experimental replications, their adoption should expand the number and variety of scientific reproducibility studies undertaken. Third, we propose educating undergraduate students to perform data analysis replications, which has scalable benefits for both the students themselves and the broader research community. In our talk we will provide details of a pilot program we created to teach undergraduates the skills necessary to conduct data analysis replications, and include a case study of the first set of students who completed this program and attempted to replicate a widely-cited social science paper on policing.

References

[1]
2018. Companion Guidelines on Replication and Reproducibility in Education Research. https://www.nsf.gov/pubs/2019/nsf19022/nsf19022.pdf
[2]
C Glenn Begley and John PA Ioannidis. 2015. Reproducibility in science: improving the standard for basic and preclinical research. Circulation research 116, 1 (2015), 116–126.
[3]
Karl Broman, Mine Cetinkaya-Rundel, Amy Nussbaum, Christopher Paciorek, Roger Peng, Daniel Turek, and Hadley Wickham. 2017. Recommendations to funding agencies for supporting reproducible research. In American Statistical Association, Vol. 2.
[4]
Katherine Button. 2018. Reboot undergraduate courses for reproducibility. Nature 561, 7723 (2018), 287–288.
[5]
Roland G. Fryer. 2019. An Empirical Analysis of Racial Differences in Police Use of Force. Journal of Political Economy 127, 3 (2019), 1210–1261. https://doi.org/10.1086/701423
[6]
Andrew Gelman and Eric Loken. 2014. The Statistical Crisis in Science. Am Sci 102, 6 (2014), 460.
[7]
Steven N. Goodman, Daniele Fanelli, and John P. A. Ioannidis. 2016. What does research reproducibility mean?Science Translational Medicine 8, 341 (2016), 341ps12–341ps12. https://doi.org/10.1126/scitranslmed.aaf5027 arXiv:https://stm.sciencemag.org/content/8/341/341ps12.full.pdf
[8]
Jake M Hofman, Amit Sharma, and Duncan J Watts. 2017. Prediction and explanation in social systems. Science 355, 6324 (2017), 486–488.
[9]
Jan H. Höffler and Thomas Kneib. 2013. Economics Needs Replication. https://www.ineteconomics.org/perspectives/blog/economics-needs-replication.
[10]
Hans IJzerman, Mark J Brandt, and Jon E Grahe. 2018. How to make replications mainstream. https://doi.org/10.31234/osf.io/rwufg
[11]
Norbert L Kerr. 1998. HARKing: Hypothesizing after the results are known. Personality and Social Psychology Review 2, 3 (1998), 196–217.
[12]
David Liu and Matthew Salganik. 2019. Successes and struggles with computational reproducibility: Lessons from the fragile families challenge. (2019).
[13]
Zacharias Maniadis, Fabio Tufano, and John A List. 2017. To replicate or not to replicate? Exploring reproducibility in economics through the lens of a model and a pilot study.
[14]
Brian A Nosek, Charles R Ebersole, Alexander C DeHaven, and David T Mellor. 2018. The preregistration revolution. Proceedings of the National Academy of Sciences 115, 11 (March 2018), 2600–2606.
[15]
Brian A Nosek and Timothy M Errington. 2019. What is replication?https://doi.org/10.31222/osf.io/u4g6t
[16]
Open Science Collaboration. 2015. Estimating the reproducibility of psychological science. Science 349, 6251 (2015). https://doi.org/10.1126/science.aac4716 arXiv:https://science.sciencemag.org/content/349/6251/aac4716.full.pdf
[17]
Prasad Patil, Roger D Peng, and Jeffrey T Leek. 2019. A visual tool for defining reproducibility and replicability. Nature human behaviour(2019), 1.
[18]
Hans E Plesser. 2018. Reproducibility vs. replicability: a brief history of a confused terminology. Frontiers in neuroinformatics 11 (2018), 76.
[19]
Patrick E Shrout and Joseph L Rodgers. 2018. Psychology, science, and knowledge construction: Broadening perspectives from the replication crisis. Annual review of psychology 69 (2018), 487–510.
[20]
Joseph P Simmons and Leif D Nelson. 2019. Data Replicada. http://datacolada.org/81.
[21]
Joseph P Simmons, Leif D Nelson, and Uri Simonsohn. 2011. False-Positive Psychology Undisclosed Flexibility in Data Collection and Analysis Allows Presenting Anything as Significant. Psychological Science 22, 11 (Oct. 2011), 0956797611417632–1366.
[22]
Uri Simonsohn, Joseph P Simmons, and Leif D Nelson. 2015. Specification Curve: Descriptive and Inferential Statistics on All Reasonable Specifications. SSRN Electronic Journal(2015).

Cited By

View all
  • (2025)Reproducible research policies and software/data management in scientific computing journals: a survey, discussion, and perspectivesFrontiers in Computer Science10.3389/fcomp.2024.14918236Online publication date: 15-Jan-2025

Index Terms

  1. Expanding the Scope of Reproducibility Research Through Data Analysis Replications
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image ACM Conferences
        WWW '20: Companion Proceedings of the Web Conference 2020
        April 2020
        854 pages
        ISBN:9781450370240
        DOI:10.1145/3366424
        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Sponsors

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        Published: 20 April 2020

        Permissions

        Request permissions for this article.

        Check for updates

        Author Tags

        1. data analysis
        2. education
        3. replication
        4. reproducibility
        5. robustness

        Qualifiers

        • Research-article
        • Research
        • Refereed limited

        Conference

        WWW '20
        Sponsor:
        WWW '20: The Web Conference 2020
        April 20 - 24, 2020
        Taipei, Taiwan

        Acceptance Rates

        Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)20
        • Downloads (Last 6 weeks)2
        Reflects downloads up to 02 Mar 2025

        Other Metrics

        Citations

        Cited By

        View all
        • (2025)Reproducible research policies and software/data management in scientific computing journals: a survey, discussion, and perspectivesFrontiers in Computer Science10.3389/fcomp.2024.14918236Online publication date: 15-Jan-2025

        View Options

        Login options

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format.

        HTML Format

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media