-
Notifications
You must be signed in to change notification settings - Fork 0
Reproducible Science
Reproducibility is a major concern in modern science, and there is somewhat of a crisis at the moment following several high-profile reports indicating that most studies are likely false and that many published results cannot be reproduced. There are many reasons for this crisis. Chief among these are:
- The use of small samples with low statistical power, which can increase the chance of false findings for several reasons (discussed in the material below)
- p-hacking; i.e., running every possible analysis to find a result with p<.05, and then reporting that result. Naturally this results in a massive multiple comparisons problem and means that the ‘significant’ results is likely false
- distorted incentives for scientists; i.e., the publish or perish culture; emphasis of prestige journals on nice, clean stories; reduce emphasis on methodological detail in these journals and so on.
In response to these trends there is a growing movement to promote reproducible science, and there are some simple things that you can do to ensure your results are robust.
-
Independent validation. There is now a large number of open MRI datasets that can allow you to replicate (most) results in an independent sample. It is worth thinking about incorporating these datasets as replication sets in your own work. Examples include:
-
1000 connectomes project/INDI: http://fcon_1000.projects.nitrc.org
-
Human connectome project: http://www.humanconnectome.org
-
Open fMRI: https://openfmri.org
-
Oasis: http://www.oasis-brains.org
-
ADNI: http://adni.loni.usc.edu
-
UK biobank: http://www.ukbiobank.ac.uk
-
NYU paediatric MRI biobank: The healthy brain network, http://fcon_1000.projects.nitrc.org/indi/cmi_healthy_brain_network/
-
Australian Schizophrenia Research Bank: http://www.schizophreniaresearch.org.au/bank/
-
An online tool for searching publicly available datasets: http://openneu.ro/metasearch/
-
A broader list of open datasets: https://github.com/cMadan/openMorph
-
Developmental datasets: see table 1 here
-
-
If you use exploratory data analysis, replicate the key finding in an independent sample, or use test-train cross-validation procedures as typically used in machine learning.
-
Get big samples.
There are many steps that should be taken to ensure that your science is reproducible. The problem can be summed up in the following quote:
"An article about computational science in a scientific publication is not the scholarship itself, it is merely advertising of the scholarship. The actual scholarship is the complete software development environment and the complete set of instructions which generated the figures.” — David Donaho
In general, you can use open code repositories such as GitHub (https://github.com) to share your code and make sure the data are accessible. It is also possible to share data using servers such as Figshare (https://figshare.com), but make sure you have ethics approval to do so first.
More details and some solutions to the problem are below:
- Great lay intro video to reproducibility problem: https://www.youtube.com/watch?v=42QuXLucH3Q
The replication crisis
-
https://simplystatistics.org/posts/2016-08-24-replication-crisis/
-
Study of biases in research: Fanelli et al., Proceedings of the National Academy of Sciences, 2017, Meta-Assessment of Bias in Science
-
Nice example of the problem in p-hacking: The Garden of Forking Paths
The problem of insufficient power
- https://garstats.wordpress.com/2017/02/04/small-sample-sizes/
- Button et al., Nature Reviews Neuroscience, 2013, Power Failure: Why Small Sample Size Undermines the Reliability of Neuroscience
- Vasishth et al., Journal of Memory and Language, 2018, The Statistical Significance Filter Leads to Overoptimistic Expectations of Replicability
- Dumas-Mallet et al., Royal Society Open Science, 2017, Low Statistical Power in Biomedical Science: A Review of Three Human Research Domains
- Loken & Gelman, Science, 2017, Measurement Error and the Replication Crisis
- The power of small N designs: Smith & Little, Psychonomic Bulletin & Review, 2018, Small Is Beautiful: In Defense of the Small-N Design
Problems with null hypothesis significance testing
-
Halsey et al., Nature Methods, 2015, The Fickle P Value Generates Irreproducible Results
-
On generalizability: Yarkoni, The Behavioral and Brain Sciences, 2020, The Generalizability Crisis
-
A modelling study on the impact of distorted incentives: Higginson & Munafò, PLOS Biology, 2016, Current Incentives for Scientists Lead to Underpowered Studies with Erroneous Conclusions
-
A tongue-in-cheek syllabus with key references on problems with current science
Best practices for data sharing in neuroimaging
-
Resources for open science: Masuzzo & Martens, Preprint, 2017, Do You Speak Open Science?
-
A signable pledge for open science lays out some core ingredients
-
Software for checking statistical results in your manuscript
-
Another course on reproducible imaging from the brain hack crew
-
An article on preregistration: Nosek et al., PNAS, 2018, The Preregistration Revolution
-
Accounting for errors in daily workflows: Strand, Psychological Methods, 2025, Error Tight
-
Inzlicht et al., SSRN Electronic Journal, 2015, News of Ego Depletion's Demise Is Premature
-
Why most published research findings are false: Ioannidis, PLoS Medicine, 2005
- 0.0 Home
- 0.1 Neuroscience fundamentals
- 0.2 Reproducible Science
- 0.3 MRI Physics, BIDS, DICOM, and data formats
- 0.4 Introduction to Diffusion MRI
- 0.5 Introduction to Functional MRI
- 0.6 Measuring functional and effective connectivity
- 0.7 Connectomics, graph theory, and complexity
- 0.8 Statistical and Mathematical Tidbits
- 0.9 Introduction to Psychopathology
- 0.10 Introduction to Genetics and Bioinformatics
- 0.11 Introduction to Programming
- 1.0 Working on the Cluster
- 2.0 Programming Languages
- 2.1 Python
- 2.2 MATLAB
- 2.3 R and RStudio
- 2.4 Programming Intro Exercises
- 2.5 git and GitHub
- 2.6 SLURM and Job Submission
- 3.0 Neuroimaging Tools and Packages
- 3.1 BIDS
- 3.2 FreeSurfer
- 3.2.1 Qdec
- 3.3 FSL
- 3.3.1 ICA-FIX
- 3.4 Connectome Workbench/wb_command
- 3.5 fMRIPrep
- 3.6 QSIPrep
- 3.7 HCP Pipeline
- 3.8 tedana
- 4.0 Quality control
- 4.1 MRIQC
- 4.2 Common Artefacts
- 4.3 T1w
- 4.4 rs-fMRI
- 5.0 Specialist Tools
- 6.0 Putting it all together