Recognition: no theorem link
Why is Regularization Underused? An Empirical Study on Trust and Adoption of Statistical Methods
Pith reviewed 2026-05-13 18:46 UTC · model grok-4.3
The pith
Survey of 606 analysts finds recommendations do not increase trust or intended use of regularization methods.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Drawing on a survey of 606 practitioners and a randomized experiment, the authors conclude that written recommendations of regularization methods have no discernible effect on trust or intended use. Adoption intentions instead depend primarily on analysts' views of implementation ease, practical advantages such as improved bias control or interpretability, and prevailing social norms.
What carries the argument
The survey instrument based on technology acceptance frameworks, combined with an embedded randomized experiment that measured trust, acceptance, and factors including perceived ease of use and social norms for regularization techniques.
If this is right
- Adoption of statistical methods depends more on perceived ease of implementation and practical benefits than on formal recommendations.
- Social norms within analyst communities act as a strong driver of intentions to use regularization.
- Promotion of new methods should prioritize demonstrating usability and tangible advantages such as bias control.
- Software interfaces that simplify regularization application could raise uptake more effectively than endorsements.
Where Pith is reading between the lines
- If self-reports align with behavior, hands-on training focused on implementation ease would likely raise actual adoption rates.
- The pattern may hold for other advanced statistical techniques beyond regularization.
- Developers could test whether interface improvements directly increase observed usage in real analyses.
Load-bearing premise
Self-reported survey measures of trust and intended use accurately reflect real-world adoption behavior of statistical methods.
What would settle it
A field study that directly observes analysts' actual code or software logs to compare real regularization usage rates against their survey-reported intentions.
Figures
read the original abstract
Statistical practice does not automatically follow methodological innovation. Regularization methods, widely advocated to reduce overfitting and stabilize inference, are readily available in modern software, but are not consistently used by data analysts. We investigate this implementation gap in a large-scale empirical study of trust in, and acceptance of, regularization techniques, based on $N = 606$ data analysts. Drawing on measurement frameworks from technology acceptance research, we survey practitioners and embed a randomized experiment to test whether written recommendation of regularization methods increases trust or intended use. We find no evidence of such an effect. Instead, adoption intentions are strongly associated with analysts' perceptions of ease of implementation and practical benefit, such as improved bias control or interpretability. Perceived social norms also emerge as a central driver. These results indicate that uptake of statistical methodology depends less on formal recommendations than on usability, perceived utility, and community practice.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript reports an empirical study of N=606 data analysts that combines a survey with an embedded randomized experiment. The central finding is that written recommendations for regularization methods produce no detectable increase in trust or intended use; instead, adoption intentions correlate strongly with perceived ease of implementation, practical benefits (e.g., bias control, interpretability), and social norms.
Significance. If the results hold after improved reporting and validation, the work provides evidence that uptake of statistical methods is driven more by usability and community practice than by formal recommendations, extending technology-acceptance frameworks to statistical methodology and offering practical guidance for methodologists seeking wider adoption.
major comments (3)
- [Methods] Methods section: the survey instrument, item wording for the trust/intended-use/ease/benefit/norms scales, response rate, and sampling frame are not described in sufficient detail to allow replication or assessment of measurement validity; the abstract and results refer to these constructs but provide no appendix or reference to validated instruments.
- [Results] Results section: the null finding for the randomized recommendation experiment is presented without effect sizes, confidence intervals, or a power analysis; without these quantities it is impossible to judge whether the study was powered to detect a practically meaningful recommendation effect.
- [Discussion] Discussion section: the claim that adoption intentions reflect real-world underuse of regularization rests entirely on self-reported scales; the manuscript contains no behavioral validation (actual code usage, follow-up task performance, or longitudinal tracking), which is load-bearing for the interpretation that perceived ease and norms, rather than recommendations, explain the implementation gap.
minor comments (2)
- [Abstract] Abstract: the reported N=606 should be accompanied by the achieved response rate and any exclusion criteria to give readers immediate context on sample representativeness.
- [Methods] Notation for the outcome scales is introduced without a clear table or appendix listing the exact Likert items or reliability coefficients (Cronbach’s α or similar).
Simulated Author's Rebuttal
We thank the referee for these constructive comments, which highlight important areas for improving clarity, statistical reporting, and interpretation. We have revised the manuscript to address each point and provide additional details below.
read point-by-point responses
-
Referee: [Methods] Methods section: the survey instrument, item wording for the trust/intended-use/ease/benefit/norms scales, response rate, and sampling frame are not described in sufficient detail to allow replication or assessment of measurement validity; the abstract and results refer to these constructs but provide no appendix or reference to validated instruments.
Authors: We agree that greater methodological transparency is required. The revised manuscript now includes a new Appendix A that reproduces the complete survey instrument, with verbatim item wording for all scales (trust, intended use, ease of implementation, practical benefits, and social norms). We have added the achieved response rate and a precise description of the sampling frame (recruitment through professional data-science forums, LinkedIn groups, and academic mailing lists). Items were adapted from established technology-acceptance instruments (Davis 1989; Venkatesh et al. 2003); we now cite these sources explicitly and note minor adaptations made for the statistical-methods context. revision: yes
-
Referee: [Results] Results section: the null finding for the randomized recommendation experiment is presented without effect sizes, confidence intervals, or a power analysis; without these quantities it is impossible to judge whether the study was powered to detect a practically meaningful recommendation effect.
Authors: We accept this criticism. The revised Results section now reports standardized effect sizes (Cohen’s d) and 95% confidence intervals for all between-condition contrasts. We have also added a post-hoc power analysis (using the observed variance and sample size) showing that the design had 80% power to detect effects as small as d = 0.23. These additions allow readers to evaluate the precision and practical significance of the null result. revision: yes
-
Referee: [Discussion] Discussion section: the claim that adoption intentions reflect real-world underuse of regularization rests entirely on self-reported scales; the manuscript contains no behavioral validation (actual code usage, follow-up task performance, or longitudinal tracking), which is load-bearing for the interpretation that perceived ease and norms, rather than recommendations, explain the implementation gap.
Authors: We acknowledge that the study relies on self-reported intentions rather than direct behavioral measures. This is a recognized limitation of survey-based technology-acceptance research. In the revised Discussion we have added an explicit limitations paragraph that (a) notes the intention–behavior gap documented in the broader literature, (b) qualifies the strength of our causal claims about real-world underuse, and (c) proposes concrete future designs (e.g., analysis of public code repositories or embedded behavioral tasks) that could provide validation. We retain the core finding that perceived ease, utility, and norms are strongly associated with stated adoption intentions, which remain theoretically and practically relevant even if they are imperfect proxies for behavior. revision: partial
Circularity Check
No significant circularity: purely empirical survey and experiment
full rationale
The paper reports results from a survey (N=606) and a randomized experiment testing recommendation effects on trust/intended use of regularization. No mathematical derivations, equations, fitted parameters presented as predictions, or self-citation chains appear in the load-bearing claims. Central findings rest on standard statistical associations from self-reported scales; these are independent of the inputs by construction and do not reduce to self-definition or renaming. The study is self-contained against external benchmarks of survey/experimental methodology.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption Self-reported survey responses validly capture perceptions, trust, and behavioral intentions
Reference graph
Works this paper leans on
-
[1]
Al-Ateeq, B., Sawan, N., Al-Hajaya, K., Altarawneh, M., & Al-Makhadmeh, A. (2022). Big Data Analytics in Auditing and the Consequences for Audit Quality: A Study Using the Technology Acceptance Model (TAM).Corporate Governance and Or- ganizational Behavior Review, 6(1), 64–78. Berger, J. O. (1985).Statistical Decision Theory and Bayesian Analysis. Springe...
-
[2]
https://doi. org/10.2307/258792 McCraw, B. W. (2015). The Nature of Epistemic Trust.Social epistemology, 29(4), 413–
-
[3]
H., Choudhury, V., & Kacmar, C
25 References Trust in Regularisation McKnight, D. H., Choudhury, V., & Kacmar, C. (2002). Developing and Validating Trust Measures for E-Commerce: An Integrative Typology.Information Systems Research, 13(3), 334–359. https://doi.org/10.1287/isre.13.3.334.81 McNeish, D. M. (2015). Using Lasso for Predictor Selection and to Assuage Overfitting: A MethodLon...
-
[4]
Sharpe, D. (2013). Why the Resistance to Statistical Innovations? Bridging the Commu- nication Gap.Psychological methods, 18(4),
work page 2013
-
[5]
StataCorp LLC. (2025).Stata Statistical Software Release 19(Software). College Station. Thiel, K. E., Baumeister, M., Krämer, N., Groll, A., Pauly, M., & Wischnewski, M. (2026). Supplementary Material of "Why Practitioners (Do Not) Use Regular- izations? An Empirical Study of Trust and Statistical Methodology Acceptance". https://doi.org/10.17877/RCTRUST-...
-
[6]
Wanner, J., Herm, L.-V., Heinrich, K., & Janiesch, C. (2022). The Effect of Transparency and Trust on Intelligent System Acceptance: Evidence from a User-Based Study. Electronic Markets, 32(4), 2079–2102. https://doi.org/10.1007/s12525-022- 00593-5 Wischnewski, M., Doebler, P., & Krämer, N. (2025, February). Development and valida- tion of the Trust in AI...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.