pith. machine review for the scientific record. sign in

arxiv: 2604.26268 · v1 · submitted 2026-04-29 · 📊 stat.AP · stat.ME

Recognition: unknown

The Difference Between "Replicable" and "Not replicable" is not Itself Scientifically Replicable

Berna Devezer, Erkan O. Buzbas

Pith reviewed 2026-05-07 12:49 UTC · model grok-4.3

classification 📊 stat.AP stat.ME
keywords replication studiesreplicability ratenon-exact replicationsheterogeneityidentifiabilityreplication crisisstatistical modelingmeta-analysis
0
0 comments X

The pith

Standard replication data with binary verdicts cannot reliably separate replicable from non-replicable scientific results.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper argues that replication studies, which aggregate binary success or failure verdicts from experiments, fail to provide a reliable way to demarcate replicable results from those that are not. This is because most replications are non-exact, with experiments differing in ways that introduce heterogeneity not captured by the usual data. The authors introduce two models for this non-exactness and demonstrate that under these models, the data cannot identify key parameters like the degree of variability, leading to understated uncertainty and apparent discriminability where none exists. A sympathetic reader would care because this undermines the foundation for declaring a replication crisis based on such studies. Common sources of heterogeneity make the problems worse, as shown in a reanalysis of existing data.

Core claim

The central discovery is that the usual data structure of replication studies—single binary verdicts per experiment—cannot support reliable demarcation between replicable and not replicable results. Under a shared latent rate model, variability in rates creates an irreducible variance floor on estimates. Under a conditionally independent rates model, the heterogeneity is not identifiable at all from the data. Thus, high- and low-replicability sequences cannot be distinguished in principle, and aggregating rates across studies lacks stable interpretation. Replicability rate is not a reliable criterion for demarcation.

What carries the argument

Two formal statistical models of non-exact replication sequences: the shared latent rate (benchmark) model, where experiments share a common random replicability rate, and the conditionally independent rates (operational) model, where each experiment has its own rate drawn from a population distribution; these models reveal that binary verdict data provides insufficient information to measure or account for between-experiment differences.

Load-bearing premise

The two statistical models of non-exactness and heterogeneity fully capture the relevant variability in actual replication studies.

What would settle it

A controlled replication dataset that collects richer continuous outcome measures per experiment and shows the estimated mean replicability rate converging without an irreducible variance floor as the number of studies grows would falsify the central claim.

Figures

Figures reproduced from arXiv: 2604.26268 by Berna Devezer, Erkan O. Buzbas.

Figure 1
Figure 1. Figure 1: Theoretical sensitivity of the estimated mean replicability rate µˆ to the degree of non-exactness ρ, under the benchmark model (equation 1). Each panel plots the 95% HDI of the sampling distribution of µˆ = X/m as a function of µ, for m ∈ {5, 50, 500} replications (light to dark). Panel A: exact replication (ρ = 0). Panels B–F: fixed non-exactness regimes ρ ∈ {0.05, 0.10, 0.15, 0.20, 0.25}. These panels d… view at source ↗
Figure 2
Figure 2. Figure 2: Effective number of independent replications me = m/[1 + (m − 1)ρ] as a function of the number of replications m, shown for six values of the intraclass correlation parameter ρ. Horizontal dotted lines mark the asymptote 1/ρ for each ρ > 0. Vertical dashed lines locate the Reproducibility Project: Psychology (RPP, m = 100) and the SCORE project (m = 274). Colored dots mark the implied me at each project’s … view at source ↗
Figure 3
Figure 3. Figure 3: Pairwise probability mass overlap between marginal posterior distributions of µ under Standard Uniform (left) and Jeffreys (right) priors on µ and ρ. Each cell shows the overlap between p(µ|x = mµi , m) and p(µ|x = mµj , m) for µi , µj ∈ {0.01, 0.12, . . . , 0.99}. The number of replications is m = 100, and for each µ the observed count of replicated results is fixed at x = mµ. The ubiquitous overlaps acro… view at source ↗
Figure 4
Figure 4. Figure 4: Conditional posterior distributions of µ with m = 100 replications and observed count x = mµtrue, under Uniform prior on µ and three fixed values of the intraclass correlation parameter: ρ = 0.05, 0.15, and 0.25. Each panel shows four posteriors corresponding to four data-generating values of µtrue ∈ {0.2, 0.4, 0.6, 0.8}, when ρ is known. Assuming ρ = 0.05, the posteriors for µtrue = 0.2 and µtrue = 0.8 ar… view at source ↗
Figure 1
Figure 1. Figure 1: Under the weakly informative prior, small- view at source ↗
Figure 5
Figure 5. Figure 5: Sampling distribution of µˆ under the benchmark model for the ML4 replication sequence (m = 17 sites, ML4 only). 95% HDI of Betabinomial distribution as a function of m, computed at ρ = 0.175, the lower bound of the 95% credible interval under Jeffreys prior ( view at source ↗
read the original abstract

Replication studies estimate the replicability rate of scientific results by aggregating binary verdicts of experiments. Exact replications are rarely attainable, so most replication sequences are non-exact. Experiments differ in ways that matter and do not share a single data-generating process. We formalize two statistical interpretations of non-exactness. In a shared latent rate (benchmark) model, experiments are exchangeable and depend on a common random replicability rate. In a conditionally independent rates (operational) model, each experiment has its own replicability rate drawn from a population distribution. Under the benchmark model, even small variability among replicability rates induces an irreducible variance floor on the estimated mean replicability rate that no amount of replication can eliminate. Under the operational model, the degree of non-exactness is not identifiable from standard replication data, because one binary verdict per experiment carries no information about between-experiment heterogeneity. Researchers cannot tell which precision regime they are in or whether high- and low-replicability sequences can be distinguished in principle. The usual data structure cannot support reliable demarcation between "replicable" and "not replicable" results and systematically understates uncertainty, making high- and low-replicability sequences appear discriminable when they are not. We show how common sources of heterogeneity amplify these problems and demonstrate practical consequences in a reanalysis of Many Labs 4. Aggregating replicability rates across heterogeneous literatures produces averages that conflate incommensurable regimes and lack a stable interpretation. Replicability rate is not a reliable demarcation criterion. The replication crisis, if there is one, cannot be established by the methods used to declare it.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 3 minor

Summary. The paper formalizes two statistical models for non-exact replication sequences— a shared latent rate (benchmark) model in which experiments are exchangeable draws from a common random replicability rate, and a conditionally independent rates (operational) model in which each experiment has its own rate drawn from a population distribution. It shows that the benchmark model induces an irreducible variance floor on the estimated mean replicability rate equal to Var(Θ), while the operational model renders the degree of heterogeneity non-identifiable from binary verdicts alone. These properties imply that standard replication data cannot reliably demarcate 'replicable' from 'not replicable' results, systematically understate uncertainty, and produce misleading distinctions between high- and low-replicability sequences. The argument is illustrated with a reanalysis of Many Labs 4 and extended to the aggregation of replicability rates across heterogeneous literatures.

Significance. If the modeling distinctions hold, the work identifies a structural limitation in the data structures routinely used to quantify replicability, showing that binary verdicts alone cannot support stable inferences about heterogeneity or replicability rates. This has direct bearing on claims about the replication crisis and on the interpretation of large-scale replication projects. The explicit derivation of the variance floor and non-identifiability results, together with the reanalysis, supplies a concrete, falsifiable basis for questioning current demarcation practices.

major comments (2)
  1. [§3] §3 (benchmark model): the derivation that the marginal variance of the sample mean converges to Var(Θ) rather than zero is a direct consequence of the hierarchical structure; however, the manuscript should state the precise regularity conditions (e.g., finite second moments of Θ) under which the asymptotic floor is guaranteed, as these are load-bearing for the 'no amount of replication can eliminate' claim.
  2. [§4] §4 (operational model): the non-identifiability result follows immediately from the fact that each experiment contributes only a single Bernoulli observation; the paper should clarify whether this remains true under mild relaxations such as known bounds on the support of the heterogeneity distribution or the availability of continuous outcome measures in a subset of replications.
minor comments (3)
  1. [reanalysis section] The reanalysis of Many Labs 4 would benefit from an explicit table or figure showing the estimated variance floor under the benchmark model versus the observed between-experiment variability.
  2. [§2] Notation for the two models (e.g., Θ vs. θ_i) is introduced clearly but could be summarized in a single display equation for quick reference.
  3. [discussion] A brief discussion of how the conclusions change if some replications provide continuous rather than binary outcomes would strengthen the practical implications.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the careful and constructive review. The comments correctly identify points where additional technical precision and scope clarification will strengthen the manuscript. We address each major comment below and have incorporated revisions as indicated.

read point-by-point responses
  1. Referee: [§3] §3 (benchmark model): the derivation that the marginal variance of the sample mean converges to Var(Θ) rather than zero is a direct consequence of the hierarchical structure; however, the manuscript should state the precise regularity conditions (e.g., finite second moments of Θ) under which the asymptotic floor is guaranteed, as these are load-bearing for the 'no amount of replication can eliminate' claim.

    Authors: We agree that the regularity conditions merit explicit statement. The revised manuscript now includes the assumption that Θ possesses finite second moments (E[Θ²] < ∞), which is the standard condition ensuring that the variance of the sample mean converges to Var(Θ) rather than zero under exchangeability. This addition does not alter the main result or the claim that an irreducible floor persists; it simply makes the technical basis transparent. revision: yes

  2. Referee: [§4] §4 (operational model): the non-identifiability result follows immediately from the fact that each experiment contributes only a single Bernoulli observation; the paper should clarify whether this remains true under mild relaxations such as known bounds on the support of the heterogeneity distribution or the availability of continuous outcome measures in a subset of replications.

    Authors: The non-identifiability result is derived specifically for the standard case of one binary verdict per experiment. With a priori known bounds on the support of the heterogeneity distribution, partial identification of the variance of rates becomes feasible in principle, though such bounds are rarely available in replication studies. Continuous outcome measures would indeed supply additional information and could relax the non-identifiability, but they fall outside the binary-verdict data structure that defines current replication protocols. The revised §4 now briefly notes these scope limitations while emphasizing that our conclusions apply to the binary data routinely collected. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation self-contained

full rationale

The paper defines two explicit hierarchical models (shared latent rate benchmark and conditionally independent rates operational) and derives their consequences for variance floors and non-identifiability directly from the model structures and standard probability results. The asymptotic variance floor equals Var(Θ) under exchangeability, and binary data yield no information on heterogeneity; both follow immediately from the stated assumptions without parameter fitting to the target replication verdicts or reduction to self-citations. No load-bearing step equates a prediction to its input by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 2 invented entities

The central claim rests on standard statistical modeling assumptions about data-generating processes for non-exact replications and the introduction of two interpretive models; no numerical free parameters are fitted to data in the described work.

axioms (2)
  • domain assumption Experiments in replication sequences are non-exact and differ in ways that matter for the data-generating process.
    Explicitly stated as the starting point for formalizing non-exactness in the abstract.
  • domain assumption Binary verdicts (replicable/not) are the only data available per experiment in standard replication studies.
    Used to derive non-identifiability in the operational model.
invented entities (2)
  • Shared latent rate (benchmark) model no independent evidence
    purpose: To represent exchangeable experiments that depend on a common random replicability rate.
    Introduced to analyze one statistical interpretation of non-exact replication.
  • Conditionally independent rates (operational) model no independent evidence
    purpose: To represent each experiment having its own replicability rate drawn from a population distribution.
    Introduced to analyze the alternative interpretation and show non-identifiability.

pith-pipeline@v0.9.0 · 5604 in / 1640 out tokens · 72350 ms · 2026-05-07T12:49:06.958520+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

19 extracted references

  1. [1]

    Raise standards for preclinical cancer research

    C Glenn Begley and Lee M Ellis. Raise standards for preclinical cancer research. Nature, 483 0 (7391): 0 531--533, 2012

  2. [2]

    Behavioural science is unlikely to change the world without a heterogeneity revolution

    Christopher J Bryan, Elizabeth Tipton, and David S Yeager. Behavioural science is unlikely to change the world without a heterogeneity revolution. Nature human behaviour, 5 0 (8): 0 980--989, 2021

  3. [3]

    Power failure: why small sample size undermines the reliability of neuroscience

    Katherine S Button, John PA Ioannidis, Claire Mokrysz, Brian A Nosek, Jonathan Flint, Emma SJ Robinson, and Marcus R Munaf \`o . Power failure: why small sample size undermines the reliability of neuroscience. Nature reviews neuroscience, 14 0 (5): 0 365--376, 2013

  4. [4]

    Buzbas, Berna Devezer, and Bert Baumgaertner

    Erkan O. Buzbas, Berna Devezer, and Bert Baumgaertner. The logical structure of experiments lays the foundation for a theory of reproducibility. Royal Society Open Science, 10 0 (3): 0 221042, 2023

  5. [5]

    Evaluating replicability of laboratory experiments in economics

    Colin F Camerer, Anna Dreber, Eskil Forsell, Teck-Hua Ho, J \"u rgen Huber, Magnus Johannesson, Michael Kirchler, Johan Almenberg, Adam Altmejd, Taizan Chan, et al. Evaluating replicability of laboratory experiments in economics. Science, 351 0 (6280): 0 1433--1436, 2016

  6. [6]

    Estimating the reproducibility of psychological science

    Open Science Collaboration. Estimating the reproducibility of psychological science. Science, 349 0 (6251): 0 aac4716, 2015

  7. [7]

    Minimum viable experiment to replicate, 2025

    Berna Devezer and Erkan Buzbas. Minimum viable experiment to replicate, 2025. URL https://philsci-archive.pitt.edu/24738/

  8. [8]

    Beyond power calculations: Assessing type s (sign) and type m (magnitude) errors

    Andrew Gelman and John Carlin. Beyond power calculations: Assessing type s (sign) and type m (magnitude) errors. Perspectives on psychological science, 9 0 (6): 0 641--651, 2014

  9. [9]

    significant

    Andrew Gelman and Hal Stern. The difference between “significant” and “not significant” is not itself statistically significant. The American Statistician, 60 0 (4): 0 328--331, 2006

  10. [10]

    Role of consciousness and accessibility of death-related thoughts in mortality salience effects

    Jeff Greenberg, Tom Pyszczynski, Sheldon Solomon, Linda Simon, and Michael Breus. Role of consciousness and accessibility of death-related thoughts in mortality salience effects. Journal of personality and social psychology, 67 0 (4): 0 627, 1994

  11. [11]

    Distribution theory for glass's estimator of effect size and related estimators

    Larry V Hedges. Distribution theory for glass's estimator of effect size and related estimators. journal of Educational Statistics, 6 0 (2): 0 107--128, 1981

  12. [12]

    Most people are not weird

    Joseph Henrich, Steven J Heine, and Ara Norenzayan. Most people are not weird. Nature, 466 0 (7302): 0 29--29, 2010

  13. [13]

    Many labs 4: Failure to replicate mortality salience effect with and without original author involvement

    Richard A Klein, Corey L Cook, Charles R Ebersole, Christine Vitiello, Brian A Nosek, Joseph Hilgard, Paul Hangsan Ahn, Abbie J Brady, Christopher R Chartier, Cody D Christopherson, et al. Many labs 4: Failure to replicate mortality salience effect with and without original author involvement. Collabra: Psychology, 8 0 (1): 0 35271, 2022

  14. [14]

    Reproducibility, replicability, and reliability

    Xiao-Li Meng. Reproducibility, replicability, and reliability. Harvard Data Science Review, 2 0 (4): 0 10, 2020

  15. [15]

    A multi-center study on the reproducibility of drug-response assays in mammalian cell lines

    Mario Niepel, Marc Hafner, Caitlin E Mills, Kartik Subramanian, Elizabeth H Williams, Mirra Chung, Benjamin Gaudio, Anne Marie Barrette, Alan D Stern, Bin Hu, et al. A multi-center study on the reproducibility of drug-response assays in mammalian cell lines. Cell systems, 9 0 (1): 0 35--48, 2019

  16. [16]

    Believe it or not: how much can we rely on published data on potential drug targets? Nature Reviews Drug Discovery, 10 0 (9): 0 712--712, 2011

    Florian Prinz, Thomas Schlange, and Khusru Asadullah. Believe it or not: how much can we rely on published data on potential drug targets? Nature Reviews Drug Discovery, 10 0 (9): 0 712--712, 2011

  17. [17]

    Investigating the replicability of the social and behavioural sciences

    Andrew H Tyner, Anna Lou Abatayo, Mason Daley, Samuel Field, Nicholas Fox, Noah A Haber, Krystal M Hahn, Melissa Kline Struhl, Brinna Mawhinney, Olivia Miske, et al. Investigating the replicability of the social and behavioural sciences. Nature, 652 0 (8108): 0 143--150, 2026

  18. [18]

    Achieving across-laboratory replicability in psychophysical scaling

    Lawrence M Ward, Michael Baumann, Graeme Moffat, Larry E Roberts, Shuji Mori, Matthew Rutledge-Taylor, and Robert L West. Achieving across-laboratory replicability in psychophysical scaling. Frontiers in Psychology, 6: 0 903, 2015

  19. [19]

    Development and validation of brief measures of positive and negative affect: the panas scales

    David Watson, Lee Anna Clark, and Auke Tellegen. Development and validation of brief measures of positive and negative affect: the panas scales. Journal of personality and social psychology, 54 0 (6): 0 1063, 1988