Threats to credible research: Where are we at?

Sarah von Grebmer zu Wolfsthurn

19/04/2026

Licence


Creative Commons Attribution 4.0

This work was originally created by Felix Schoenbrodt under a CC-BY 4.0 Creative Commons Attribution 4.0 International License. This current work by Sarah von Grebmer zu Wolfsthurn, Malika Ihle and Felix Schoenbrodt is licensed under a CC-BY-SA 4.0 Creative Commons Attribution 4.0 International SA License. It permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited. If you remix, transform, or build upon the material, you must distribute your contributions under the same license as the original.

Contribution statement


Creator: Von Grebmer zu Wolfsthurn, Sarah (orcid logo 0000-0002-6413-3895)

Reviewer: Schönbrodt, Felix (orcid logo0000-0002-8282-3910)

Consultant: Ihle, Malika (orcid logo 0000-0002-3242-5981)

Prerequisites

Important

Before completing this submodule, please carefully read about the prerequisites.

Prerequisite Description Link/Where to find it
UNESCO Recommendations on Open Science Recommended reading: pp 6-19 Download Link

Before we start - survey time

Based on your experience so far, how would you currently rate your trust in published scientific findings on a scale from 1 - 5? (1 = not trusting any of the findings, 2 = trusting only some findings, 3 = trusting about half of the findings, 4 = trusting the majority of the findings, 5 = trusting all findings)

  1. 1

  2. 2

  3. 3

  4. 4

  5. 5

Based on your experience so far, do you currently see any challenges in research?

Display Wordcloud answer.

Based on your experience so far, which concepts to you connect to research more broadly?

Display Wordcloud answer.

What is your level of familiarity with Open Research practices in general (e.g., basic concepts, terminology, or tools)?

  1. I am unfamiliar with the concept of Open Research practices.

  2. I have heard of them but I would not know how they apply to my work.

  3. I have basic understanding and experience with Open Research practices in my own work/research/studies.

  4. I am very familiar with Open Research practices and routinely apply them in my daily work/research/study routines.

Discussion of survey results


What do we see in the results?

Where are we at?

Previously:

  • Introduction to this module

Up next:

  • Getting started: Current threats to credible research
  • Diving into the solution

Covered in this session

  • Research and the research cycle
  • “Trust” in research and what it means
  • Replicability and reproducibilty
  • Biases, inconsistencies and mistakes in research
  • Working towards a solution: Open Research

Learning goals

By the end of this session, learners will be able to:

  • Define and distinguish key terms related to research, including research cycle, trust in research, replicability, reproducibility, and open research
  • Recognize different types of challenges in research: research biases, statistical insecurities, errors and where they can arise from
  • Analyze research scenarios to identify potential research biases and questionable research practices
  • Reflect on the current threats for research and on the need for alternative approaches to conducting research

Key terms and definitions - TO DO

  • Aim: Introduce key terms and definitions that students will come across throughout the session.

  • Key Term 1: Definition

  • Key Term 2: Definition

  • Key Term 3: Definition

Practical exercise 1

Task: Collect your thoughts with your neighbor.



What does “research” mean to you?

What is “research”?


“Research refers to a careful, well-defined (or redefined), objective, and systematic method of search for knowledge, or formulation of a theory that is driven by inquisitiveness for that which is unknown and useful on a particular aspect so as to make an original contribution to expand the existing knowledge base. Research involves the formulation of hypothesis or proposition of solutions, data analysis, and deductions; and ascertaining whether the conclusions fit the hypothesis. Research is a process of creating, or formulating knowledge that does not yet exist.



Deb, D., Dey, R., & Balas, V. E. (2018). Introduction: What is research? In Engineering Research Methodology: A Practical Insight for Researchers (pp. 1-7). Singapore: Springer Singapore.

The research cycle

A  diagram illustrating the scientific, with six stages arranged clockwise: Observation/Question, Research topic area, Hypothesis, Test with experiment, Analyze data, and Report Conclusions connected by arrows to show the ongoing, iterative research process.

www.phdcomics.com

The research cycle

A circular diagram illustrating the design-based research cycle, with four stages arranged horizontally: Observation, formulation hypothesis, test hypothesis with experiment, establish theory based on repeated validation of results connected by arrows to show the ongoing, iterative research process.

DBR English greyscaler (Design-based research cycle)” by Sarah Zloklikovits, licensed under CC BY 4.0 — Wikimedia Commons.

Practical exercise 2

Task: For the following “observation”, map out the individual steps of the research cycle:


“I wonder what happens to pasta when I cook it in unsalted water?”

Can we trust in research?

What is (public) “trust” in research?

  • Society trusts that scientific research results are an honest and accurate reflection of a researcher’s work.
    (Committee on Science, Engineering and Public Policy 2009: ix)
  • The public must be able to trust the science and scientific process informing public policy decisions.
    (Obama 2009)

Resnik, D. B. (2011). Scientific research and the public trust. Science and engineering ethics, 17(3), 399-409.

Measuring trust in research

  • Via social surveys: “How much do you trust researchers to tell the truth?” (Example: UK)

Bar chart showing levels of public trust in different professions in the UK (2024). Professions are ranked from most to least trusted based on the percentage of people who say they trust them to tell the truth. Nurses rank highest (94%), followed by engineers (90%), doctors (88%), and teachers and professors (both around 85%). Mid-ranking professions include police, judges, and business leaders. Lower levels of trust are shown for journalists (around 27%), advertising executives (18%), government ministers (15%), and politicians (11%), which is the lowest. Overall, the chart shows high trust in healthcare and technical professions and much lower trust in political roles.

Sources: Veracity Index 2024, Ipsos; Seyd, B. (2025). What is trust (in science and scientists) and is it in crisis?. Current Opinion in Psychology, 67. https://doi.org/10.1016/j.copsyc.2025.102201

Take aways

  • Seyd (2025):
    • Trust in scientists remains relatively high in many countries
    • Overall trust levels in research remain stable
    • But: public skepticism about scientists’ integrity and transparency

Seyd, B. (2025). What is trust (in science and scientists) and is it in crisis?. Current Opinion in Psychology, 67. https://doi.org/10.1016/j.copsyc.2025.102201

Public trust in research

Add alt text

Source: Wissenschaft im Dialog/Verian. The original text of the questionnaire as well as all result tables are available online via the following link: www.sciencebarometer.com.

What is (academic) “trust” in research?

  • No consensus on definition from the perspective of researchers

  • Trust is essential for effective collaboration among researchers (includes co-authorship, peer review, data sharing, replication, teaching, mentoring etc.)

  • Scientists reading published research trust that the work was conducted as described, that all relevant methodological details are disclosed, and that the data have not been fabricated or falsified

Resnik, D. B. (2011). Scientific research and the public trust. Science and engineering ethics, 17(3), 399-409.

Practical exercise 3

As a researcher, what can you do to make your pasta experiment trustworthy?

Tip

Think about your approach when formulating your hypothesis, when conducting your experiment, when analyzing your data, when writing up your findings; but also about potential confounding variables or hurdles you could encounter during the research process.

Replicability and reproducibility

Feature Replicability Reproducibility
Definition Ability to repeat an experiment using the same methods and obtain the same results Ability to obtain consistent results using the original data and code
Focus aka repeating the experiment and collecting new data aka re-analyzing the original data with the original code etc.
Materials Same experiment setup, protocols, conditions etc. Original data, analysis scripts, code etc.
In practise Running the same psychological experiment with new participants Running the published analysis on the original dataset

Replicability and reproducibility

Additional exercises

Want to practice how to distinguish the two? Skip to the end of the slides for additional exercises on replicability vs. reproducibility.

Replicability and reproducibility across disciplines

Begley, C. G., & Ellis, L. M. (2012); Camerer et al (2016); Chang & Li (2015); Cova et al. (2018); Open Science Collaboration (2015); Social Science: Combined sample of systematically sampled projects (RPP, SSRP, EERP); Prinz, F., Schlange, T., & Asadullah, K. (2011); Protzko et al. (2023)

Replicability in the social and behavioural sciences

Tyner et al. (2026):

  • Authors attempted replications of 274 claims of positive results from 164 quantitative papers (published between 2009 and 2018)
  • ~55% of positive claims replicated successfully
Discipline Replication attempts (successful / total) Percentage successful
Business 17 / 36 47.2%
Economics 10.2 / 24 42.5%
Education 8.2 / 13 63.1%
Political science 7.8 / 15 52.0%
Psychology 28.4 / 58 49.0%
Sociology 9.2 / 18 51.1%

Tyner, S. K., Abatayo, A. L., Dayley, M., et al. (2025). Investigating the replicability of the social and behavioural sciences. Nature. https://doi.org/10.1038/s41586-025-10078-y

Psychology: The Reproducibility Project (2015)

  • Large-scale replication project:
    • Close/exact replications of 100 experimental and correlational studies from 3 different psychological journals
    • Reproducibility evaluated based on effect sizes, p-values, subjective assessment of replication teams
    • Contacted original study authors when necessary

Psychology: The Reproducibility Project (2015)

  • Large-scale replication project:
    • Close/exact replications of 100 experimental and correlational studies from 3 different psychological journals
    • Reproducibility evaluated based on effect sizes, p-values, subjective assessment of replication teams
    • Contacted original study authors when necessary

What did they find?

Large portion of replications produced weaker evidence for the original findings despite using materials provided by the original authors.

Open Science Collaboration. (2015). Estimating the reproducibility of psychological science. Science, 349(6251), aac4716.

Not a recent issue

An image with te titles of journal articles reporting about the replicability and reproducibility crisis across difference disciplines.

Adapted from Dr. Malika Ihle: https: https://osf.io/u3znx

Why did not more studies replicate?

human error
lack of resources
cognitive bias
publication pressure
??
lack of statistical knowledge

Publication pressure

Why published papers are useful:

  • Getting a job

  • Being awarded grant money for your research

  • Being visible in the respective research field

Publication pressure

Why published papers are useful:

  • Getting a job

  • Being awarded grant money for your research

  • Being visible in the respective research field

Consequence: Rat race culture

Researchers try to publish as much as they can and to outperform their peers (Schmidt et al., 2021).

Career relevance of publishing

  • Survey among N = 1453 psychology researchers, 66% were actually members of a professorship hiring committee
Actual (not desired) relevance in professorship hiring committees Rank
Number of peer-reviewed publications 1
Fit of research profile to the hiring department 2
Quality of research talks 3
Number of publications 4
Volume of acquired third party funding 5
Number of first authorships 6

Abele-Brehm, A. E., & Bühner, M. (2016). Wer soll die Professur bekommen? Psychologische Rundschau, 67(4), 250–261. http://doi.org/10.1026/0033-3042/a000335

How do I get lots of publications?

significant results

Fanelli, D. (2010). “Positive” results increase down the hierarchy of the sciences. PLOS ONE, 5, e10068. https://doi.org/10.1371/journal.pone.0010068

Publication bias

“If my study works, I can publish it. If it does not, let’s hide it the drawer.

Publication bias happens when studies with positive or significant results are much more likely to be published than studies with negative or non-significant results.

Song, F., Hooper, L., & Loke, Y. K. (2013). Publication bias: what is it? How do we measure it? How do we avoid it?. Open Access Journal of Clinical Trials, 71-81. https://doi.org/10.2147/OAJCT.S34419

Example publication bias

Turner study

Turner, E. H., Matthews, A. M., Linardatos, E., Tell, R. A., & Rosenthal, R. (2008). Selective publication of antidepressant trials and its influence on apparent efficacy. New England Journal of Medicine, 358(3), 252-260. https://doi.org/10.1056/NEJMsa065779

More than just publication bias

De Vries, Y. A., Roest, A. M., de Jonge, P., Cuijpers, P., Munafò, M. R., & Bastiaansen, J. A. (2018). The cumulative effect of reporting and citation biases on the apparent efficacy of treatments: the case of depression. Psychological Medicine, 48(15), 2453–2455. https://doi.org/10.1017/S0033291718001873

  1. All conducted studies
  2. Study publication bias: non-publication of an entire study
  3. Outcome reporting bias: non-publication of negative outcomes within a published article or switching the status of (non-significant) primary and (significant) secondary outcomes
  4. Spin: authors conclude that the treatment is effective despite non-significant results on the primary outcome
  5. Citation bias: Studies with positive results receive more citations than negative studies

“Thought so already” bias

confirming your own beliefs

“I knew it” bias

confirming your own beliefs

Practical exercise 4

Task: Match the bias to its description.

Bias Description
1. Confirmation bias A. Non-publication of negative outcomes within a paper, or switching non-significant primary outcomes with significant secondary ones.
2. Spin B. Studies with positive results receive more citations than negative studies.
3. Study publication bias C. After learning the outcome, believing “I knew it all along.”
4. Hindsight bias D. Non-publication of an entire study (e.g., trials with null results never submitted).
5. Outcome reporting bias E. Tendency to seek or interpret information in ways that confirm existing beliefs.
6. Citation bias F. Authors conclude the treatment is effective despite non-significant primary outcomes.

Pre-break survey

Based on what you learnt so far, how would you currently rate your trust in published scientific findings on a scale from 1 - 5? (1 = not trusting any of the findings, 2 = trusting only some findings, 3 = trusting about half of the findings, 4 = trusting the majority of the findings, 5 = trusting all findings)

  1. 1

  2. 2

  3. 3

  4. 4

  5. 5

What does replicability in research mean?

  1. Obtaining the same results using the original dataset and code

  2. Obtaining consistent results when a new study collects new data using the same methods

  3. Publishing results in more than one journal

  4. Repeating the statistical analysis multiple times

What is publication bias in research?

  1. The tendency for journals to publish studies only from well-known researchers

  2. The requirement that all published studies must be peer-reviewed

  3. The practice of publishing the same study in multiple journals

  4. The tendency for studies with positive results to be published more often than studies with non-significant or negative results

Break! 15 minutes

Post-break survey discussion


What do we see in the results?

Accidental p-hacking

“P-hack… What now?”

p-hacking and questionble research practises

Excursus: Null-Hypothesis-Significance Testing

Example: Does the new drug “SniffStop” work better to decrease flu symptoms compared to the existing drug “CoughAway”?

  • H₀ = “the new drug is not better than the existing one.”
  • Collect data from experiments and perform a statistical analysis to see if the evidence is strong enough to reject H₀
  • P-value = how likely it is to see the results you got if H₀ is true.
    • A p-value of 0.05 means: “There’s a 5% chance of seeing these results (or more extreme) if the null hypothesis is true.

Important

A p-value of 0.05 means that we accept a 5% chance that our results came about by pure luck (if the results were accidental, we would speak of a false positive result)

“Hack” 1: Add lots of outcome variables

  • For two outcome variables: False positive rate increases from 5% to 9.5%

  • For five outcome variables: False positive rate increases from 5% to 41%

p-hacking and questionble research practises

“Hack” 2: Run as many comparisons as possible

  • Run as many different comparisons on different outcomes, subgroups, time windows etc. as you can
  • Only report the comparisons that produced a statistically significant result

O’Boyle, E. H., Banks, G. C., & Gonzalez-Mulé, E. (2017). The Chrysalis Effect: How ugly initial results metamorphosize into beautiful articles. Journal of Management, 43(2), 376–399. https://doi.org/10.1177/0149206314527133

“Hack” 3: Stop collecting data whenever you found what you were looking for

  • Collect an initial sample, analyze the results, add participants if the results are not significant
  • Stop when significance is found
    • One analysis: α = 5%
    • Two analyses: α = 11%
    • But with enough “just looking” can be pushed to 100%!

Armitage, P., McPherson, C. K., & Rowe, B. C. (1969). Repeated significance tests on accumulating data. Journal of the Royal Statistical Society. Series A (General), 132, 235–244.

“Hack” 4: Drop participants you do not like

  • Selectively exclude data/ outliers after seeing the results until the results are satisfying (aka until significance has been reached)

“Hack” 5: HARK-ing

  • Hypothesizing After Results are Known = presenting an exploratory finding to match a hypothesis that was created only after analysing the results
p-hacking and questionable research practises

Combining these “hacks”

p-hacking and questionble research practises
  • Combining some of these hacks (aka questionable research practices) can raise false positive rates from 5% to > 50%!
  • The logic of the p-value is therefore corrupted and “renders the reported p-values essentially uninterpretable.

Stefan, A. M., & Schönbrodt, F. D. (2023). Big little lies: A compendium and simulation of p -hacking strategies. Royal Society Open Science, 10(2), 220346. https://doi.org/10.1098/rsos.220346

A note to remember

Warning

The so-called “hacks” on the past few slides represent questionable research practices. Do not try at home.

Practical exercise 5

Scenario: You are reviewing a study examining whether drinking white tea improves short-term memory, where the researchers report:

  • Hypothesis: White tea improves memory test scores.

  • Sample size: 28 participants per group (tea-drinkers vs. water-only-drinkers)

  • Results:

    • Effect was “stronger in women” (p = 0.049)
    • Effect “even stronger when excluding two outliers” (p = 0.044)
    • No effect in men (p = 0.31)
    • Reaction time difference significant (p = 0.046)
  • Conclusion: The results show that white tea reliably improves cognitive performance.

Which potential p-hacking strategies are at play here?

Practical exercise 5

Scenario: You are reviewing a study examining whether drinking white tea improves short-term memory, where the researchers report:

  • Hypothesis: White tea improves memory test scores.

  • Sample size: 28 participants per group (tea-drinkers vs. water-only-drinkers)

  • Results:

    • Effect was “stronger in women” (p = 0.049)
    • Effect “even stronger when excluding two outliers” (p = 0.044)
    • No effect in men (p = 0.31)
    • Reaction time difference significant (p = 0.046)
  • Conclusion: The results show that white tea reliably improves cognitive performance.

The “real” scientific method

diagram of the real scientific method by phdcomics.com to illustrate the QRPs and biased nature inherent to the research cycle.

Does this actually happen in real life?

Come on now. Surely not..?

p-hacking and questionble research practises

John, L. K., Loewenstein, G., & Prelec, D. (2012). Measuring the prevalence of questionable research practices with incentives for truth telling. Psychological science, 23(5), 524-532.

.. and across fields?

  • Survey among 6,813 academic researchers in The Netherlands: Self-reported prevalence of fabrication and falsification in the last 3 years
QRPs in psychology

Gopalakrishna, G., ter Riet, G., Vink, G., Stoop, I., Wicherts, J. M., & Bouter, L. M. (2022). Prevalence of questionable research practices, research misconduct and their potential explanatory factors: A survey among academic researchers in The Netherlands. PLOS ONE, 17(2), e0263023. https://doi.org/10.1371/journal.pone.0263023

.. and across fields?

  • Survey among 6,813 academic researchers in The Netherlands: Self-reported prevalence of fabrication and falsification in the last 3 years
QRPs in psychology

Gopalakrishna, G., ter Riet, G., Vink, G., Stoop, I., Wicherts, J. M., & Bouter, L. M. (2022). Prevalence of questionable research practices, research misconduct and their potential explanatory factors: A survey among academic researchers in The Netherlands. PLOS ONE, 17(2), e0263023. https://doi.org/10.1371/journal.pone.0263023

(Un)Intentional?

  • Intentional?
    • “Evil researcher” who only cares about his/her career and not at all about truth-seeking?
    • We urge the social science community to redefine p-hacking as a series of deceptive research practices rather than ones that are merely questionable.(Craig et al., 2020)
  • Unintentional?
    • Lack of education/knowledge?
    • Wrong/uncritical standards of the field?
    • Pushed by supervisors, reviewers, or editors? ➙ http://bulliedintobadscience.org/
    • Simply being human?

Human errors and honest mistakes

“90% Excel-Gate”

Lessons learnt

Important

The most important point of the story: The original authors shared their raw data, which made it possible to correct the honest mistake!

Statistical errors

QRPs in psychology
  • Reproducible analysis code and open data required at submission - “in-house checking” in review process

  • 54% of all submissions had results in the paper that did not match the computed results from the code

    • Wrong signs, wrong labeling of regression coefficients, errors in sample sizes, wrong descriptive stats

Eubank, N. (2016). Lessons from a decade of replications at the quarterly journal of political science. PS: Political Science & Politics, 49(2), 273-276 https://doi.org/10.1017/S1049096516000196

Statistical inconsistencies

QRPs in psychology
  • 16,695 scanned papers with statcheck tool (Nuijten et al., 2015)
  • 50% of papers contain statistical inconsistencies
  • 13% contained strong errors (i.e., where the statistical conclusion changes).
  • Numerical results of less than 30% of papers can be reproduced (Crüwell et al., 2022)

Nuijten, M. B., Hartgerink, C. H., Van Assen, M. A., Epskamp, S., & Wicherts, J. M. (2016). The prevalence of statistical reporting errors in psychology (1985–2013). Behavior research methods, 48(4), 1205-1226. https://doi.org/10.3758/s13428-015-0664-2

What does this all mean?

Bias + (accidental) p-hacking + human (honest) mistakes = untrustworthy research findings?

Note

  • Published findings across fields to be viewed with caution?
  • “We know” –> “We think we know”?
  • More research to verify existing “truths”?

A romanticized idea of research?

An image describing a romantic view of research: The smartest heads in the world immerse themselves into a research topic for years. In that process, they become the experts, nobody knows more about that topic.The boundaries of knowledge have been pushed forward. When the researchers are confident in their findings, they publish them in the best scientific journals, with the highest standards of quality, rigor, and integrity.

Going even further: The perpetuating role of AI

Emsley, R. (2023). ChatGPT: These are not hallucinations - they’re fabrications and falsifications. Schizophrenia, 9(1), 52. Lautrup, A. D., Hyrup, T., Schneider-Kamp, A., Dahl, M., Lindholt, J. S., & Schneider-Kamp, P. (2023). Heart-to-heart with ChatGPT: the impact of patients consulting AI for cardiovascular health advice. Open Heart, 10(2). Shekar, S., Pataranutaporn, P., Sarabu, C., Cecchi, G. A., & Maes, P. (2025). People Overtrust AI-Generated Medical Advice despite Low Accuracy. NEJM AI, 2(6), AIoa2300015.

Now what?

crossroads

This image was taken from the Geograph project collection. The copyright on this image is owned by Chris Martin and is licensed for reuse under the Creative Commons Attribution-ShareAlike 2.0 license.

A new way of doing research


Open Research
aka
A scientific framework for the 21. century

To be continued …

Reflection activity



One-minute paper: Imagine you would have to explain the current challenges in research you heard about today to a friend. Write down what you would say to them.

Take-home message

What are you taking away from today?

Take-home message

What are you taking away from today?

Remember: There are solutions!

Research is not “doomed” - on the contrary. More on this in the next session!

Thanks!

See you next class :)

Additional exercises: Replicability and reproducibility

Decide whether each scenario in the following slides is an example of reproducibility or replicability.

Scenario 1



A computational neuroscientist reruns a published fMRI analysis using the original dataset and Python scripts to verify the reported brain activation patterns.

Scenario 2



An environmental scientist repeats a field experiment on soil nutrient levels using the same sampling protocol at a different site.

Scenario 3



A linguist reanalyzes a corpus of historical texts using the same annotation guidelines and code to verify reported patterns of syntactic structures.

Scenario 4



A psychology lab replicates a social behavior experiment using new participants from a different cultural background.

Psychology: The Replication Database by FORRT

  • Collects replication results across different psychological fields
  • Provide detailed overview of the original findings and the replication outcomes
An image of the landing page of the FORRT Replication database

FORRT Replication Database (https://forrt-replications.shinyapps.io/fred_explorer/)

References

Anderson, L. W., & Krathwohl, D. R. (2001). A taxonomy for learning, teaching, and assessing : A revision of Bloom’s taxonomy of educational objectives : Complete edition. Addison Wesley Longman, Inc.
Bloom, B. S. (1956). The 1955 Normative Study of the Tests of General Educational Development. The School Review, 64(3), 110–124. https://doi.org/10.1086/442296
Craig, R., Cox, A., Tourish, D., & Thorpe, A. (2020). Using retracted journal articles in psychology to understand research misconduct in the social sciences: What is to be done? Research Policy, 49(4), 103930. https://doi.org/10.1016/j.respol.2020.103930
Crüwell, S., Apthorp, D., Baker, B. J., Colling, L. J., Elson, M., Geiger, S. J., Lobentanzer, S., Monéger, J., Patterson, A., Schwarzkopf, D. S., Zaneva, M., & Brown, N. J. L. (2022). What’s in a badge? A computational reproducibility investigation of the Open Data badge policy in one Issue of Psychological Science. PsyArXiv. https://doi.org/10.31234/osf.io/729qt
Schmidt, R., Curry, S., & Hatch, A. (2021). Creating SPACE to evolve academic assessment. eLife, 10, e70929. https://doi.org/10.7554/eLife.70929

Additional slides TÓ REMOVE?

What published research can be replicated/reproduced? REMOVE

Field Success Failure
OSC (2015) – Psychology 36% 64%
Chang & Li (2015) – Economics (67 papers, 29 papers replicated) 43% 57%
Camerer 2016 – Econ laboratory 61% 39%
Camerer combined Social Sci 62% 38%
Begley & Ellis (2012) – Cancer Research 11% 89%
Prinz et al. (2011) – Pharmaceutical research 35% 65%
Cova et al. (2018) – x-philosophy 70% 30%
Protzko et al. (2023) – Social 86% 14%

Why should we trust researchers?

snakesman analogy

…right? MOVE

chasing significance

Decide where to add - maybe not needed?

Ideal scenario: Balancing the desire to stay truthful to research with the necessity to publish?