Abstract
In statistics, samples are drawn from a population in a data generating process (DGP). Standard errors measure the uncertainty in sample estimates of population parameters. In science, evidence is generated to test hypotheses in an evidence generating process (EGP). We claim that EGP variation across researchers adds uncertainty: non-standard errors. To study them, we let 164 teams test six hypotheses on the same sample. We find that non-standard errors are sizeable, on par with standard errors. Their size (i) co-varies only weakly with team merits, reproducibility, or peer rating, (ii) declines significantly after peer-feedback, and (iii) is underestimated by participants.
Original language | English |
---|---|
Place of Publication | London |
Publisher | CEPR Press (Centre for Economic Policy Research) |
Number of pages | 56 |
DOIs | |
Publication status | Published - Nov 2021 |
Austrian Fields of Science 2012
- 502009 Corporate finance
- 502014 Innovation research