The statistical significance filter leads to overoptimistic expectations of replicability

It is well-known in statistics (e.g., Gelman & Carlin, 2014) that treating a result as publishable just because the p-value is less than 0.05 leads to overoptimistic expectations of replicability. These effects get published, leading to an overconfident belief in replicability. We demonstrate th...

Full description

Bibliographic Details
Main Authors: Gelman, A. (Author), Jäger, L.A (Author), Mertzen, D. (Author), Vasishth, S. (Author)
Format: Article
Language:English
Published: Academic Press Inc. 2018
Subjects:
Online Access:View Fulltext in Publisher
Description
Summary:It is well-known in statistics (e.g., Gelman & Carlin, 2014) that treating a result as publishable just because the p-value is less than 0.05 leads to overoptimistic expectations of replicability. These effects get published, leading to an overconfident belief in replicability. We demonstrate the adverse consequences of this statistical significance filter by conducting seven direct replication attempts (268 participants in total) of a recent paper (Levy & Keller, 2013). We show that the published claims are so noisy that even non-significant results are fully compatible with them. We also demonstrate the contrast between such small-sample studies and a larger-sample study; the latter generally yields a less noisy estimate but also a smaller effect magnitude, which looks less compelling but is more realistic. We reiterate several suggestions from the methodology literature for improving current practices. © 2018 Elsevier Inc.
ISBN:0749596X (ISSN)
DOI:10.1016/j.jml.2018.07.004