Key Moments
The Reproducibility Crisis
Key Moments
Psychology faces a reproducibility crisis due to publication bias, p-hacking, low statistical power, and harking. Solutions include registered reports and open data.
Key Insights
The reproducibility crisis means many published results, particularly in psychology, cannot be replicated.
Key factors contributing to the crisis include publication bias (favoring positive results), p-hacking (manipulating data analysis), insufficient statistical power (small sample sizes), and harking (hypothesizing after results are known).
Pharmaceutical companies are concerned as they cannot reproduce foundational scientific results for drug development.
Registered reports, where study methods are peer-reviewed and accepted before data collection, offer a promising solution by decoupling publication from results.
Social media has empowered early-career researchers to voice concerns and advocate for scientific reform.
Addressing the crisis requires changes in incentive structures within academia, greater awareness from funders, and public engagement that fosters trust rather than distrust.
IDENTIFYING THE REPRODUCIBILITY CRISIS
The reproducibility crisis refers to the growing concern that many scientific results, especially within psychology, are not being replicated. Initially dismissed by some as a minor statistical issue, alarming findings from studies attempting to reproduce published work revealed that only 30-40% of results were reproducible. This prompted widespread alarm and a deeper investigation into the causes and potential solutions across various scientific disciplines, including the biomedical sciences.
THE FOUR HORSEMEN OF THE REPRODUCIBILITY APOCALYPSE
Dorothy Bishop outlines four primary contributors to the crisis: 1. Publication bias, where positive and exciting results are preferentially published, distorting the scientific literature and creating a 'big file drawer' of null or negative findings. 2. P-hacking, the practice of analyzing data in multiple ways until a statistically significant result (p < 0.05) is found, which can lead to spurious correlations. 3. Low statistical power, stemming from studies with sample sizes too small to reliably detect the often-modest effect sizes common in psychology. 4. Harking, or hypothesizing after the results are known, where researchers present findings as pre-planned hypotheses after discovering them through exploratory data analysis.
CHALLENGES IN SAMPLE SIZE AND DATA ANALYSIS
Obtaining large sample sizes can be challenging in fields that study specific populations or require expensive data collection methods, such as brain imaging or rare patient groups. This often necessitates collaboration to pool resources and increase sample sizes. Furthermore, the flexibility in data analysis techniques can lead to incorrect statistical significance if each analytical approach is treated as a new attempt to find a result, underscoring the importance of pre-specified analysis plans.
REGISTERED REPORTS AS A PUBLICATION SOLUTION
A promising solution gaining traction is the 'registered report' publication model. In this system, a study's introduction and methods are peer-reviewed and accepted in principle before data collection begins. This pre-review process specifies the analysis plan, and if the researchers adhere to it, the paper is guaranteed publication. This approach effectively mitigates publication bias, p-hacking, and harking by divorcing the publication decision from the study's outcomes.
THE EVOLVING ROLE OF SOCIAL MEDIA AND EDUCATION
Social media has become a crucial platform for early-career researchers to voice concerns about reproducibility issues, fostering a more militant approach to demanding scientific rigor. New educational methods, such as using simulated data to teach statistical concepts, are also being introduced to help researchers understand the implications of P-hacking and insufficient power. These educational shifts aim to equip future scientists with better tools and awareness from the outset.
INCENTIVE STRUCTURES AND PUBLIC TRUST
Addressing the deep-seated issues requires reforming academic incentive structures, so that universities and funders prioritize research credibility over publication in high-impact journals. Funders, motivated by a desire for reliable research, are increasingly pushing for change. While openness about the crisis is vital, it must be balanced against the risk of it being 'weaponized' by those who wish to undermine scientific consensus. Ultimately, improving the self-correcting mechanisms of science will build public trust.
Mentioned in This Episode
●Software & Apps
●Companies
●Organizations
●Books
●Concepts
●People Referenced
Common Questions
The reproducibility crisis refers to the recent observation, particularly over the last 10-15 years, that many published results in psychology (and other scientific fields) are difficult or impossible to replicate. A major study found that only 30-40% of published findings could be reproduced.
Topics
Mentioned in this video
A scientific journal where Dorothy Bishop published a commentary on the reproducibility crisis.
Medical Research Council, a UK research council that supported the Academy of Medical Sciences symposium and motivates change in research credibility.
A journal that has agreed to offer Registered Reports as a publication option.
University where Chris Chambers is based, a proponent of Registered Reports.
Biotechnology and Biological Sciences Research Council, a UK research council that supported the Academy of Medical Sciences symposium and motivates change in research credibility.
An institution that held a symposium on the reproducibility crisis, particularly concerning biomedical sciences and the pharmaceutical industry.
A funder that supported the Academy of Medical Sciences symposium on reproducibility and motivates change by not wanting to fund non-credible research.
Developed the Registered Reports publication model.
Professor of psychology at the University of Oxford and author of a Nature magazine commentary on the reproducibility crisis.
Colleague of Dorothy Bishop at Oxford University in anthropology who is heading a bid for university-wide funding to coordinate reproducibility initiatives across disciplines.
A social media platform where junior scientists can quickly voice concerns and draw attention to issues in scientific publications, fostering a more proactive community.
Expressed concern over the reproducibility crisis as it impacted their ability to build on fundamental research from experimental labs.
Analyzing data in multiple ways or collecting data on many variables and only reporting those yielding statistically significant results (p < 0.05), distorting the literature.
The issue of conducting studies with sample sizes that are too small to reliably detect genuine effects, leading to missed findings or erroneous conclusions.
The tendency to publish positive, exciting results more easily than null or negative results, leading to a distorted literature.
Hypothesizing After the Results are Known; a practice where researchers analyze data, find an interesting result, and then present it as if it were the original hypothesis.
A publication model where studies are reviewed and accepted in-principle based on their introduction and methods before data collection, mitigating issues like p-hacking and publication bias.
The threshold (commonly p < 0.05) used to determine if a result is likely due to chance or represents a real effect, often misunderstood.
More from Sabine Hossenfelder
View all 65 summaries
7 minBreakthrough In Data Storage Could Store Your Photos for 10000 Years
7 minThe Simulation Hypothesis Gets Scientific Backing
7 minSurprise! Milky Way Might Not Have a Black Hole After All
7 minThe First Moon Landing Wasn’t Apollo — And We Just Found It
Found this useful? Build your knowledge library
Get AI-powered summaries of any YouTube video, podcast, or article in seconds. Save them to your personal pods and access them anytime.
Try Summify free