Improving Openness and Reproducibility of Scientific Research

Slides:



Advertisements
Similar presentations
Good data practices Jelte M. Wicherts 1. 2 Source: Wicherts, J. M. (2011). Psychology must learn a lesson from fraud case. Nature, 480, 7.
Advertisements

Improving Integrity, Transparency, and Reproducibility Through Connection of the Scholarly Workflow Andrew Sallans Partnerships Lead Center for Open Science.
Open Science Framework: Supporting the research worflow Brian Nosek University of Virginia.
Chapter One: The Science of Psychology
“Knowing Revisited” And that’s how we can move toward really knowing something: Richard Feynman on the Scientific Method.
Chapter One: The Science of Psychology. Ways to Acquire Knowledge Tenacity Tenacity Refers to the continued presentation of a particular bit of information.
Main issues Effect-size ratio Development of protocols and improvement of designs Research workforce and stakeholders Reproducibility practices and reward.
The Psychologist as Detective, 4e by Smith/Davis © 2007 Pearson Education Chapter One: The Science of Psychology.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
Scientific Utopia: Improving Openness and Reproducibility Brian Nosek University of Virginia Center for Open Science.
Practical Steps for Increasing Openness and Reproducibility Courtney Soderberg Statistical and Methodological Consultant Center for Open Science.
Webinar on increasing openness and reproducibility April Clyburne-Sherin Reproducible Research Evangelist
April Center for Open Fostering openness, integrity, and reproducibility of scientific research.
Practical Steps for Increasing Openness and Reproducibility Courtney Soderberg Statistical and Methodological Consultant Center for Open Science.
Smith/Davis (c) 2005 Prentice Hall Chapter One The Science of Psychology PowerPoint Presentation created by Dr. Susan R. Burns Morningside College.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
David Mellor, PhD Project Manager at Improving Openness and Reproducibility of Scientific Research.
Sara Bowman Center for Open Science | Promoting, Supporting, and Incentivizing Openness in Scientific Research.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
Open Science Framework Jeffrey Center for Open Science | University of Virginia.
Sara Bowman Center for Open Science | Promoting, Supporting, and Incentivizing Openness in Scientific Research.
Dr.V.Jaiganesh Professor
Improving Openness and Reproducibility of Scientific Research
F1000: Open for science Hollydawn Murray
David Preregistration David
David Mellor Building infrastructure to connect, preserve, speed up, and improve scholarship David Mellor
Safer science: making psychological science more Transparent & replicable Simine Vazire UC Davis.
Increasing openness, reproducibility, and prediction in social science research My general substantive interest in the gap between values and practices.
Transparency increases credibility and relevance of research
Scholarly Workflow: Federal Prototype and Preprints
Working with Scholarly Articles
Improving Openness and Reproducibility of Scientific Research
Improving Openness and Reproducibility of Scientific Research
Principles of Quantitative Research
Shifting the research culture toward openness and reproducibility
Center for Open Science: Practical Steps for Increasing Openness
Jarek Nabrzyski Director, Center for Research Computing
Experimental Psychology
Three points 1. Scientists’ Conflict of Interest 2
Improving Openness and Reproducibility of Scientific Research
David Mellor Building infrastructure to connect, preserve, speed up, and improve scholarship David Mellor
Psi Chi’s Network for International Collaborative Exchange (NICE)
Open Science Framework
An Open Science Framework for Managing and Sharing Research Workflows
AF1: Thinking Scientifically
Open Science Framework
Achieving Open Science
Data Sharing Now and in the Future
Transparency increases the credibility and relevance of research
Preregistration on the Open Science Framework
A Framework for Managing and Sharing Research Workflow
Research & Writing in CJ
Reinventing Scholarly Communication by Separating Publication From Evaluation Brian Nosek University of Virginia -- Center for Open Science
Shifting incentives from getting it published to getting it right
The Basics of Literature Reviews
Incentives for a more #openscience
Improving Openness and Reproducibility of Scientific Research
Modularity and Interoperability
Study Pre-Registration
Disrupting Scholarly Communication
Reading Research Papers-A Basic Guide to Critical Analysis
How to Read A Scientific Paper
Unit 1 Lesson 2 Scientific Investigations
Unit 1 Lesson 2 Scientific Investigations
School of Psychology, Cardiff University
The Scientific Method.
Judy MIELKE, PhD. Taylor & Francis
Badges to Acknowledge Open Practices
Process of the Scientific Method
Open Science & Reproducibility
Presentation transcript:

Improving Openness and Reproducibility of Scientific Research Brian Nosek University of Virginia -- Center for Open Science http://briannosek.com/ -- http://cos.io/ My general substantive interest in the gap between values and practices. The work that I am discussing today is a practical application of this interest to the gap between scientific values and practices. In particular, how can I best advance knowledge and my career at the same time? Challenges I face when working to advance scientific knowledge and my career at the same time. And, how my scientific practices can be adapted to meet my scientific values.

Norms Counternorms Communality Secrecy Open sharing Closed Communality – open sharing with colleagues; Secrecy Universalism – research evaluated only on its merit; Particularism – research evaluated by reputation/past productivity Disinterestedness – scientists motivated by knowledge and discovery, not by personal gain; self-interestedness – treat science as a competition with other scientists Organized skepticism – consider all new evidence, theory, data, even if it contradicts one’s prior work/point-of-view; organized dogmatism – invest career in promoting one’s own most important findings, theories, innovations Quality – seek quality contributions; Quantity – seek high volume

Norms Counternorms Communality Universalism Secrecy Particularlism Open sharing Universalism Evaluate research on own merit Secrecy Closed Particularlism Evaluate research by reputation Communality – open sharing with colleagues; Secrecy Universalism – research evaluated only on its merit; Particularism – research evaluated by reputation/past productivity Disinterestedness – scientists motivated by knowledge and discovery, not by personal gain; self-interestedness – treat science as a competition with other scientists Organized skepticism – consider all new evidence, theory, data, even if it contradicts one’s prior work/point-of-view; organized dogmatism – invest career in promoting one’s own most important findings, theories, innovations Quality – seek quality contributions; Quantity – seek high volume

Norms Counternorms Communality Universalism Disinterestedness Secrecy Open sharing Universalism Evaluate research on own merit Disinterestedness Motivated by knowledge and discovery Secrecy Closed Particularlism Evaluate research by reputation Self-interestedness Treat science as a competition Communality – open sharing with colleagues; Secrecy Universalism – research evaluated only on its merit; Particularism – research evaluated by reputation/past productivity Disinterestedness – scientists motivated by knowledge and discovery, not by personal gain; self-interestedness – treat science as a competition with other scientists Organized skepticism – consider all new evidence, theory, data, even if it contradicts one’s prior work/point-of-view; organized dogmatism – invest career in promoting one’s own most important findings, theories, innovations Quality – seek quality contributions; Quantity – seek high volume

Norms Counternorms Communality Universalism Disinterestedness Open sharing Universalism Evaluate research on own merit Disinterestedness Motivated by knowledge and discovery Organized skepticism Consider all new evidence, even against one’s prior work Secrecy Closed Particularlism Evaluate research by reputation Self-interestedness Treat science as a competition Organized dogmatism Invest career promoting one’s own theories, findings Communality – open sharing with colleagues; Secrecy Universalism – research evaluated only on its merit; Particularism – research evaluated by reputation/past productivity Disinterestedness – scientists motivated by knowledge and discovery, not by personal gain; self-interestedness – treat science as a competition with other scientists Organized skepticism – consider all new evidence, theory, data, even if it contradicts one’s prior work/point-of-view; organized dogmatism – invest career in promoting one’s own most important findings, theories, innovations Quality – seek quality contributions; Quantity – seek high volume

Norms Counternorms Communality Universalism Disinterestedness Open sharing Universalism Evaluate research on own merit Disinterestedness Motivated by knowledge and discovery Organized skepticism Consider all new evidence, even against one’s prior work Quality Secrecy Closed Particularlism Evaluate research by reputation Self-interestedness Treat science as a competition Organized dogmatism Invest career promoting one’s own theories, findings Quantity Communality – open sharing with colleagues; Secrecy Universalism – research evaluated only on its merit; Particularism – research evaluated by reputation/past productivity Disinterestedness – scientists motivated by knowledge and discovery, not by personal gain; self-interestedness – treat science as a competition with other scientists Organized skepticism – consider all new evidence, theory, data, even if it contradicts one’s prior work/point-of-view; organized dogmatism – invest career in promoting one’s own most important findings, theories, innovations Quality – seek quality contributions; Quantity – seek high volume

Anderson, Martinson, & DeVries, 2007

Incentives for individual success are focused on getting it published, not getting it right Nosek, Spies, & Motyl, 2012

Problems Low power Flexibility in analysis Selective reporting Ignoring nulls Lack of replication Examples from: Button et al – Neuroscience Ioannidis – why most results are false (Medicine) GWAS Biology Two possibilities are that the percentage of positive results is inflated because negative results are much less likely to be published, and that we are pursuing our analysis freedoms to produce positive results that are not really there. These would lead to an inflation of false-positive results in the published literature. Some evidence from bio-medical research suggests that this is occurring. Two different industrial laboratories attempted to replicate 40 or 50 basic science studies that showed positive evidence for markers for new cancer treatments or other issues in medicine. They did not select at random. Instead, they picked studies considered landmark findings. The success rates for replication were about 25% in one study and about 10% in the other. Further, some of the findings they could not replicate had spurred large literatures of hundreds of articles following up on the finding and its implications, but never having tested whether the evidence for the original finding was solid. This is a massive waste of resources. Across the sciences, evidence like this has spurred lots of discussion and proposed actions to improve research efficiency and avoid the massive waste of resources linked to erroneous results getting in and staying in the literature, and about the culture of scientific practices that is rewarding publishing, perhaps at the expense of knowledge building. There have been a variety of suggestions for what to do. For example, the Nature article on the right suggests that publishing standards should be increased for basic science research. [It is not in my interest to replicate – myself or others – to evaluate validity and improve precision in effect estimates (redundant). Replication is worth next to zero (Makel data on published replications; motivated to not call it replication; novelty is supreme – zero “error checking”; not in my interest to check my work, and not in your interest to check my work (let’s just each do our own thing and get rewarded for that) Irreproducible results will get in and stay in the literature (examples from bio-med). Prinz and Begley articles (make sure to summarize accurately) The Nature article by folks in bio-medicine is great. The solution they offer is a popular one in commentators from the other sciences -- raise publishing standards. Sterling, 1959; Cohen, 1962; Lykken, 1968; Tukey, 1969; Greenwald, 1975; Meehl, 1978; Rosenthal, 1979

Figure credit: fivethirtyeight.com Silberzahn et al., 2015

http://compare-trials.org/

Median effect size (d) = .29 % p < .05 = 63% Reported Tests (122) Median p-value = .02 Median effect size (d) = .29 % p < .05 = 63% Unreported Tests (147) Median p-value = .35 Median effect size (d) = .13 % p < .05 = 23% We find that about 40% of studies fail to fully report all experimental conditions and about 70% of studies do not report all outcome variables included in the questionnaire. Reported effect sizes are about twice as large as unreported effect sizes and are about 3 times more likely to be statistically significant. N = 32 studies in psychology Unreported tests (147) Median p-value = .35 Median d = .13 % significant = 23% Reported tests (N = 122) Median p = .02 Median d = .29 % sig p<.05 = 63% Franco, Malhotra, & Simonovits, 2015, SPPS

Positive Result Rate dropped from 57% to 8% after preregistration required.

Barriers Perceived norms (Anderson, Martinson, & DeVries, 2007) Motivated reasoning (Kunda, 1990) Minimal accountability (Lerner & Tetlock, 1999) I am busy (Me & You, 2016) We can understand the nature of the challenge with existing psychological theory. For example: 1. The goals and rewards of publishing are immediate and concrete; the rewards of getting it right are distal and abstract (Trope & Liberman) 2. I have beliefs, ideologies, and achievement motivations that influence how I interpret and report my research (motivated reasoning; Kunda, 1990). And, even if I am trying to resist this motivated reasoning. I may simply be unable to detect it in myself, even when I can see those biases in others. 3. And, what biases might influence me. Well, pick your favorite. My favorite in this context is the hindsight bias. 4. What’s more is we face these potential biases in a context of minimal accountability. What you know of my laboratory work is only what you get in the published report. … 5. Finally, even if I am prepared to accept that I have these biases and am motivated to address them so that I can get it right. I am busy. So are you. If I introduce a whole bunch of new things that I must now do to check and correct for my biases, I will kill my productivity and that of my collaborators. So, the incentives lead me to think that my best course of action is to just to the best I can and hope that I’m doing it okay.

MEANS REWARDS Novel, Positive, Clean Transparency, Reproducibility Outcomes Process MEANS Research Content Data and Materials Publication REWARDS

Signals: Making Behaviors Visible Promotes Adoption Badges Open Data Open Materials Preregistration Psychological Science (Jan 2014) Kidwell et al., 2016

40% 30% % Articles reporting data available in repository 20% 10% 0%

PREREGISTRATION Context of Justification Confirmation Data independent Hypothesis testing Context of Discovery Exploration Data contingent Hypothesis generating p-values interpretable p-values NOT interpretable PREREGISTRATION Presenting exploratory as confirmatory increases publishability of results at the cost of credibility of results Study 1 Study 1 Study 2

Positive Result Rate dropped from 57% to 8% after preregistration required.

Are you okay with receiving treatment based on clinical trials that were not preregistered? “oh but clinical trials are important” Why would I waste my time on something that isn’t important enough to be worth doing the best that I can do it?

Preregistration Challenge http://cos.io/prereg

Registered Reports PEER REVIEW Design Collect & Analyze Report Publish Review of intro and methods prior to data collection; published regardless of outcome Beauty vs. accuracy of reporting Publishing negative results Conducting replications Peer review focuses on quality of methods http://osf.io/8mpji, Committee Chair: Chris Chambers

MEANS REWARDS Novel, Positive, Clean Transparency, Reproducibility Outcomes Process MEANS Research Content Data and Materials Publication REWARDS

Mundane and Big Challenges for Reproducibility Forgetting Losing materials and data

http://osf.io

Technological innovation to open all of science What is coming? Technological innovation to open all of science

Open Access Open Data Open Workflows Outcomes Content Process Improving scientific ecosystem Make outcomes more accessible Make research content more accessible Make research process more accessible

osf.io OSF Application Framework journals registries preprint servers Workflow Authentication File Storage File Rendering Meta-database Integrations Search SHARE Data osf.io journals registries preprint servers grants management OSF is actually an application framework It is a public good and scholarly commons It supports the interface you see if you google OSF Blog engine, Slide sharing (osf.io/meetings) Workflow integration Authentication File storage File rendering Database Metadata/Annotations/Commenting External service integrations Search SHARE Data peer review services curation, annotation

Application Framework Branded Services OSF Preprints: SocArXiv, PsyArXiv, engrXiv [+ bioRxiv, chemRxiv, arXiv, …] Views / Interfaces OSF, OSF Preprints, OSF Registries, OSF Institutions, OSF Meetings [+ many more possible] Application Framework JamDB, API V2, Ember OSF, Ops/Deploy/Security, Keen/Analytics, Search, Documentation, Django, Localization, SHARE

http://osf.io/preprints/

Application Framework JamDB, API V2, Ember OSF, Ops/Deploy/Security, Keen/Analytics, Search, Documentation, Django, Localization, SHARE

Breaking tyranny of the publication Separating publication from evaluation Nosek & Bar-Anan, 2012

Community of Preprint servers Journals Moderation (pre-peer review) Commenting / Discussions (post-peer review)

Technology to enable change Training to enact change Incentives to embrace change Improving scientific ecosystem

UNIVERSITIES ecosystem PUBLISHING FUNDERS SOCIETIES

What can you do? Try OSF, http://osf.io/ Prereg Challenge, http://cos.io/prereg/ Share a preprint, http://osf.io/preprints/ Editors: Badges, Registered Reports, TOP Departments: OSF-Reproducibility workshops, hiring and promotion criteria Individuals: COS Ambassador Email: Support@cos.io or nosek@virginia.edu

http://improvingpsych.org/

So is there a crisis? Nature survey. Who are these respondents?

And why are they saying that – pervasive challenge across disciplines in reproducing prior results. A fundamental feature of science is that scientific claims gain credibility compared to other kinds of claims based on the potential to independently reproduce the evidence. But, of course, whether there is a reproducibility crisis or not isn’t really an a relevant question. The real question is whether there are cultural norms and practices that are undermining the efficiency and effectiveness of knowledge building, and are there ways that we can do better?

Conclusion Psychology is not under threat, it is leading the way to more open, reproducible science Start with conclusion…. Why such focus on psych? We are getting the attention not because we have bigger problems, but because we are bothering to face them and we have productive ideas and are taking steps to deal with them. Think about how much time you spend not caring about what is happening in physics, chemistry, earth sciences, and biology. Turns out, that is just about the same amount of time that physicists, chemists, earth scientists, and biologists spend not caring about us. That’s why the “reproducibility crisis” is actually an opportunity for us, not a threat. Other areas of science will care about us more and more not when we solve our challenges, but when we solve theirs. And, that’s already underway…

What can you do? Try out OSF, http://osf.io/ Prereg Challenge, http://cos.io/prereg/ Share a preprint, http://osf.io/preprints/ Join SIPS, http://improvingpsych.org/ These slides are shared at: http://osf.io/bq4kn [take a picture of this slide] Email: Support@cos.io or nosek@virginia.edu