Key Moments

TL;DR

Psychology faces a reproducibility crisis due to publication bias, p-hacking, low statistical power, and harking. Solutions include registered reports and open data.

Key Insights

1

The reproducibility crisis means many published results, particularly in psychology, cannot be replicated.

2

Key factors contributing to the crisis include publication bias (favoring positive results), p-hacking (manipulating data analysis), insufficient statistical power (small sample sizes), and harking (hypothesizing after results are known).

3

Pharmaceutical companies are concerned as they cannot reproduce foundational scientific results for drug development.

4

Registered reports, where study methods are peer-reviewed and accepted before data collection, offer a promising solution by decoupling publication from results.

5

Social media has empowered early-career researchers to voice concerns and advocate for scientific reform.

6

Addressing the crisis requires changes in incentive structures within academia, greater awareness from funders, and public engagement that fosters trust rather than distrust.

IDENTIFYING THE REPRODUCIBILITY CRISIS

The reproducibility crisis refers to the growing concern that many scientific results, especially within psychology, are not being replicated. Initially dismissed by some as a minor statistical issue, alarming findings from studies attempting to reproduce published work revealed that only 30-40% of results were reproducible. This prompted widespread alarm and a deeper investigation into the causes and potential solutions across various scientific disciplines, including the biomedical sciences.

THE FOUR HORSEMEN OF THE REPRODUCIBILITY APOCALYPSE

Dorothy Bishop outlines four primary contributors to the crisis: 1. Publication bias, where positive and exciting results are preferentially published, distorting the scientific literature and creating a 'big file drawer' of null or negative findings. 2. P-hacking, the practice of analyzing data in multiple ways until a statistically significant result (p < 0.05) is found, which can lead to spurious correlations. 3. Low statistical power, stemming from studies with sample sizes too small to reliably detect the often-modest effect sizes common in psychology. 4. Harking, or hypothesizing after the results are known, where researchers present findings as pre-planned hypotheses after discovering them through exploratory data analysis.

CHALLENGES IN SAMPLE SIZE AND DATA ANALYSIS

Obtaining large sample sizes can be challenging in fields that study specific populations or require expensive data collection methods, such as brain imaging or rare patient groups. This often necessitates collaboration to pool resources and increase sample sizes. Furthermore, the flexibility in data analysis techniques can lead to incorrect statistical significance if each analytical approach is treated as a new attempt to find a result, underscoring the importance of pre-specified analysis plans.

REGISTERED REPORTS AS A PUBLICATION SOLUTION

A promising solution gaining traction is the 'registered report' publication model. In this system, a study's introduction and methods are peer-reviewed and accepted in principle before data collection begins. This pre-review process specifies the analysis plan, and if the researchers adhere to it, the paper is guaranteed publication. This approach effectively mitigates publication bias, p-hacking, and harking by divorcing the publication decision from the study's outcomes.

THE EVOLVING ROLE OF SOCIAL MEDIA AND EDUCATION

Social media has become a crucial platform for early-career researchers to voice concerns about reproducibility issues, fostering a more militant approach to demanding scientific rigor. New educational methods, such as using simulated data to teach statistical concepts, are also being introduced to help researchers understand the implications of P-hacking and insufficient power. These educational shifts aim to equip future scientists with better tools and awareness from the outset.

INCENTIVE STRUCTURES AND PUBLIC TRUST

Addressing the deep-seated issues requires reforming academic incentive structures, so that universities and funders prioritize research credibility over publication in high-impact journals. Funders, motivated by a desire for reliable research, are increasingly pushing for change. While openness about the crisis is vital, it must be balanced against the risk of it being 'weaponized' by those who wish to undermine scientific consensus. Ultimately, improving the self-correcting mechanisms of science will build public trust.

Common Questions

The reproducibility crisis refers to the recent observation, particularly over the last 10-15 years, that many published results in psychology (and other scientific fields) are difficult or impossible to replicate. A major study found that only 30-40% of published findings could be reproduced.

Topics

Mentioned in this video

More from Sabine Hossenfelder

View all 65 summaries

Found this useful? Build your knowledge library

Get AI-powered summaries of any YouTube video, podcast, or article in seconds. Save them to your personal pods and access them anytime.

Try Summify free