BERKELEY INITIATIVE FOR TRANSPARENCY IN THE SOCIAL SCIENCES Garret Christensen, Research Fellow BITSS and Berkeley Institute for Data Science.

Slides:



Advertisements
Similar presentations
Progress on Registration in Political Science Jeremy M. Weinstein (on behalf of Experiments in Governance and Politics – EGAP)
Advertisements

Improving Integrity, Transparency, and Reproducibility Through Connection of the Scholarly Workflow Andrew Sallans Partnerships Lead Center for Open Science.
Registration in political science reflections and proposals Macartan Humphreys.
Pre-analysis plans Module 8.3. Recap on statistics If we find a result is significant at the 5% level, what does this mean? – there is a 5% or less probability.
Evidenced Based Practice; Systematic Reviews; Critiquing Research
Cohort Studies Hanna E. Bloomfield, MD, MPH Professor of Medicine Associate Chief of Staff, Research Minneapolis VA Medical Center.
©2007 Prentice Hall Organizational Behavior: An Introduction to Your Life in Organizations Chapter 19 OB is for Life.
The New Statistics: Why & How Corey Mackenzie, Ph.D., C. Psych.
Systematic Reviews.
Research Utilization in Nursing Chapter 21
Systematic reviews to support public policy: An overview Jeff Valentine University of Louisville AfrEA – NONIE – 3ie Cairo.
Evidence-Based Medicine: What does it really mean? Sports Medicine Rounds November 7, 2007.
Ensuring rigour in qualitative research CPWF Training Workshop, November 2010.
1 Judy Hewitt, PhD On Detail to Office of Extramural Research National Institutes of Health May 18, 2015 Center for Scientific Review Advisory Council.
Systematic Reviews Michael Chaiton Tobacco and Health: From Cells to Society September 24, 2014.
Systematic Review: Interpreting Results and Identifying Gaps October 17, 2012.
Pre-Analysis Plans: Pros and Cons of Limiting Discretion in the Analysis of RCTs Donald P. Green Columbia University.
Professor Phillipa Hay Centre for Health Research, School of Medicine.
Stephanie Wykstra March 7-8, 2016 Research Transparency Workshop Maanzoni Lodge, Athi River, Kenya Research reliability: The problems and the way forward.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
Scientific Utopia: Improving Openness and Reproducibility Brian Nosek University of Virginia Center for Open Science.
Practical Steps for Increasing Openness and Reproducibility Courtney Soderberg Statistical and Methodological Consultant Center for Open Science.
Webinar on increasing openness and reproducibility April Clyburne-Sherin Reproducible Research Evangelist
April Center for Open Fostering openness, integrity, and reproducibility of scientific research.
Practical Steps for Increasing Openness and Reproducibility Courtney Soderberg Statistical and Methodological Consultant Center for Open Science.
Leamer-Rosenthal Prizes  Introduction of Leaders in Education  Edward Miguel, UC Berkeley  Introduction of Emerging Researchers  Social Science – Edward.
Scientific Utopia: I. Improving Scientific Communication Brian Nosek University of Virginia Center for Open Science.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
David Mellor, PhD Project Manager at Improving Openness and Reproducibility of Scientific Research.
Sara Bowman Center for Open Science | Promoting, Supporting, and Incentivizing Openness in Scientific Research.
Systematic Reviews of Evidence Introduction & Applications AEA 2014 Claire Morgan Senior Research Associate, WestEd.
The Reproducible Research Advantage Why + how to make your research more reproducible Presentation for the Center for Open Science June 17, 2015 April.
Brian Nosek University of Virginia -- Center for Open Science -- Improving Openness.
Open Science Framework Jeffrey Center for Open Science | University of Virginia.
Sara Bowman Center for Open Science | Promoting, Supporting, and Incentivizing Openness in Scientific Research.
BERKELEY INITIATIVE FOR TRANSPARENCY IN THE SOCIAL SCIENCES Garret Christensen, Research Fellow BITSS and Berkeley Institute for Data Science.
David Preregistration David
Evaluating the effectiveness of open science practices
David Mellor Building infrastructure to connect, preserve, speed up, and improve scholarship David Mellor
Safer science: making psychological science more Transparent & replicable Simine Vazire UC Davis.
Increasing openness, reproducibility, and prediction in social science research My general substantive interest in the gap between values and practices.
What is Open Science and How do I do it?
Critically Appraising a Medical Journal Article
The New Statistics: Why & How Corey Mackenzie, Ph.D., C. Psych
Improving Openness and Reproducibility of Scientific Research
Improving Openness and Reproducibility of Scientific Research
Reproducibility Project: Psychology A Discussion
Shifting the research culture toward openness and reproducibility
The Research Design Continuum
Lorne Campbell University of Western Ontario
Three points 1. Scientists’ Conflict of Interest 2
David Mellor Building infrastructure to connect, preserve, speed up, and improve scholarship David Mellor
Preregistration Challenge
Achieving Open Science
Transparency increases the credibility and relevance of research
Preregistration on the Open Science Framework
A Framework for Managing and Sharing Research Workflow
The Reproducible Research Advantage
Reinventing Scholarly Communication by Separating Publication From Evaluation Brian Nosek University of Virginia -- Center for Open Science
Shifting incentives from getting it published to getting it right
Clinical Study Results Publication
Research Transparency and Reproducibility Training (RT2)
Incentives for a more #openscience
Methods Hour: Preregistration
Study Pre-Registration
Disrupting Scholarly Communication
Challenges for Journals: Encouraging Sound Science
School of Psychology, Cardiff University
Chapter 4 Summary.
Open Science & Reproducibility
Presentation transcript:

BERKELEY INITIATIVE FOR TRANSPARENCY IN THE SOCIAL SCIENCES Garret Christensen, Research Fellow BITSS and Berkeley Institute for Data Science UC San Diego, February 2016

Why transparency?  Public policy and private decisions are based on evaluation of past events (i.e. research).  So research can affect millions of lives.  But what is a “good” evaluation?  Credibility  Legitimacy

Scientific values Merton 1942  Communality Open Sharing of knowledge 2.Universalism Anyone can make a claim 3.Disinterestedness “Truth” as motivation (≠COI) 4.Organized Skepticism Peer review, replication

Why we worry… Anderson, Martinson, DeVries 2007

A response:

Ecosystem for Open Science

Why we worry… What we’re finding: Weak academic norms can distort the body of evidence.  Publication bias (“file drawer” problem)  P-hacking  Non-disclosure  Selective reporting  Failure to replicate

Publication Bias “File drawer problem”

Publication Bias  Status quo: Null results are not as “interesting.”  What if you find no relationship between a school intervention and test scores? (in a well-designed study…)  It’s less likely to get published, so null results are hidden.  How do we know? Rosenthal 1979:  Published: 3 published studies, all showing a positive effect…  Hidden: A few unpublished studies showing null effect  The significance of positive findings is now in question!

In social sciences… Franco, Malhotra, Simonovits 2014

Turner et al. [2008] ClinicalTrials.gov In medicine…

P-Hacking  Scientists want to test hypotheses  i.e. look for relationships among variables (schooling, test scores)  Observed relationships should be statistically significant  Minimize the likelihood that an observed relationship is actually a false discovery  Common norm: p-value < 0.05 But null results not “interesting”... So incentive is to look for (or report) the positive effects, even if they’re false discoveries

Turner et al. [2008] In economics… Brodeur et al Data: 50,000 tests published in AER, JPE, QJE ( )

In sociology… Gerber and Malhotra 2008

In political science… Gerber and Malhotra 2008

Solution: Registries Prospectively register hypotheses in a public database. “Paper trail” to solve the “File Drawer” problem. Differentiate confirmatory hypothesis testing from exploratory.  Medicine & Public Health: clinicaltrials.govclinicaltrials.gov  Economics: AEA registry: socialscienceregistry.orgsocialscienceregistry.org  Political Science: EGAP Registry: egap.org/design-registrationegap.org/design-registration  Development: 3IE Registry: ridie.3ieimpact.orgridie.3ieimpact.org  Open Science Framework: Open Questions:  How best to promote registration? Nudges, incentives (Registered Reports, Badges), requirements (journal standards), penalties?  How to adjust for observational (non-experimental) work?

Solution: Registries  $1,000,000 Pre-Reg Challenge

Solution: Results-blind Review  Review study design before seeing the results, issue in- principal acceptance.  “Registered Reports”  Chris Chambers, Psychology  20+ journals   Forthcoming Comparative Political Studies  Findley, et al. (2016)

Non-disclosure  To evaluate the evidentiary quality of research, we need complete reporting of methods and results….  Challenge: limited real estate in journals  Challenge: heterogeneous reporting  Challenge: perverse incentives  It’s impossible to replicate or validate findings if methods are not disclosed.

Solution: Standards Nosek et al, 2015 Science

Organizational Efforts  TOP Guidelines:  DA-RT Guidelines:  Psych Science Guidelines: Checklists for reporting excluded data, manipulations, outcome measures, sample size. Inspired by grass-roots psychdisclosure.org. Psych Science Guidelines psychdisclosure.org

Grass Roots Efforts  21 word solution in Nelson, Simmons and Simonsohn (2012): “We report how we determined our sample size, all data exclusions (if any), all manipulations, and all measures in the study.”  Peer Reviewer Openness Initiative: “We suggest that beginning January 1, 2017, reviewers make open practices a pre-condition for more comprehensive review. ” Peer Reviewer Openness Initiative

Selective Reporting  Problem: Cherry-picking & fishing for results  Can result from vested interests, perverse incentives… You can tell many stories with any data set… Example: Casey, Glennerster and Miguel (2012, QJE)

Solution: Pre-specify 1.Define hypotheses 2.Identify all outcomes to be measured 3.Specify statistical models, techniques, tests (# obs, sub- group analyses, control variables, inclusion/exclusion rules, corrections, etc)  Pre-Analysis Plans: Written up just like a publication. Stored in registries, can be embargoed.  Open Questions: will it stifle creativity? Could “thinking ahead” improve the quality of research?  Unanticipated benefit: Protect your work from political interests!

Reduce Cherry-Picking Casey, Glennerster, Miguel (QJE 2012)

Failure to replicate “Reproducibility is just collaboration with people you don’t know, including yourself next week”— Philip Stark, UC Berkeley “Economists treat replication the way teenagers treat chastity - as an ideal to be professed but not to be practised.”—Daniel Hamermesh, UT Austin

Why we care  Identifies fraud, human error  Confirms earlier findings (bolsters evidence base)

Replication Resources  Replication Wiki:   Large-scale Replication Efforts  Reproducibility Project: Psychology Reproducibility Project: Psychology  Many Labs Many Labs

Replication Resources  Data/Code Repositories:  Dataverse (IQSS)  ICPSR  Open Science Framework  GitHub  New ICMJE Policy?  Jan 20, 2016 

Replication Standards Replications need to be subject to rigorous peer review (no “second-tier” standards) Could they be pre-registered as well?

Reproducibility The Reproducibility Project: Psychology was a crowdsourced effort to estimate the reproducibility of a sample of 100 studies from the literature. Science (Aug 28, 2015): “Ninety-seven percent of original studies had statistically significant results. Thirty-six percent of replications had statistically significant results; 47% of original effect sizes were in the 95% confidence interval of the replication effect size; 39% of effects were subjectively rated to have replicated the original result”

Many Labs

Why we worry… Some Solutions…  Publication bias  Pre-registration  Non-disclosure  Reporting standards  P-hacking  Pre-specification  Failure to replicate  Open data/materials, Many Labs

What does this mean? Pre-register study and pre-specify hypotheses, protocols & analyses Carry out pre-specified analyses; document process & pivots Report all findings; disclose all analyses; share all data & materials BEFOREDURINGAFTER In practice:

Report everything another researcher would need to replicate your research: Literate programming Version control Dynamic documents Follow consensus reporting standards

RAISING AWARENESS about systematic weaknesses in current research practices FOSTERING ADOPTION of approaches that best promote scientific integrity IDENTIFYING STRATEGIES and tools for increasing transparency and reproducibility BITSS Focus

Raising Awareness

 Social Media:  Publications  Best Practices Manual  Textbook, MOOC  Sessions at conferences: AEA/ASSA, APSA, MozFest  BITSS Summer Institute (June)  BITSS Annual Meeting (December) Raising Awareness

 Tools  Open Science Framework: osf.io  Registries: AEA, EGAP, 3ie, Clinicaltrials.gov  Coursework  Syllabi  Slide decks Identifying Strategies

 Annual Summer Institute in Research Transparency (  Consulting with COS (  Meta-research grants (  Leamer-Rosenthal Prizes for Open Social Science ( Fostering Adoption

Year Two: RFP Out Now  New methods to improve the transparency and credibility of research?  Systematic uses of existing data (innovation in meta-analysis) to produce credible knowledge?  Understanding research culture and adoption of new norms?  More info: SSMART Grants

Year Two: Fall 2016

bitss.org cega.org