Funded through the ESRC’s Researcher Development Initiative Department of Education, University of Oxford Session 2.3 – Publication bias.

Slides:



Advertisements
Similar presentations
Hypothesis Testing Steps in Hypothesis Testing:
Advertisements

EVAL 6970: Meta-Analysis Vote Counting, The Sign Test, Power, Publication Bias, and Outliers Dr. Chris L. S. Coryn Spring 2011.
Specifying an Econometric Equation and Specification Error
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
The Campbell Collaborationwww.campbellcollaboration.org Moderator analyses: Categorical models and Meta-regression Terri Pigott, C2 Methods Editor & co-Chair.
Estimation and Reporting of Heterogeneity of Treatment Effects in Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare.
Introduction to Meta-Analysis Joseph Stevens, Ph.D., University of Oregon (541) , © Stevens 2006.
Overview of Lecture Parametric Analysis is used for
Chapter 11: Inference for Distributions
Practical Meta-Analysis -- D. B. Wilson 1 Practical Meta-Analysis David B. Wilson.
Today Concepts underlying inferential statistics
Chapter 14 Inferential Data Analysis
Richard M. Jacobs, OSA, Ph.D.
Simple Linear Regression Analysis
Sample Size Determination Ziad Taib March 7, 2014.
Multiple Linear Regression A method for analyzing the effects of several predictor variables concurrently. - Simultaneously - Stepwise Minimizing the squared.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Are the results valid? Was the validity of the included studies appraised?
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Inference for regression - Simple linear regression
Chapter 13: Inference in Regression
Statistical Techniques I
Determining Sample Size
Regression Method.
Daniel Acuña Outline What is it? Statistical significance, sample size, hypothesis support and publication Evidence for publication bias: Due.
Funded through the ESRC’s Researcher Development Initiative
Statistics for Education Research Lecture 5 Tests on Two Means: Two Independent Samples Independent-Sample t Tests Instructor: Dr. Tung-hsien He
The Examination of Residuals. The residuals are defined as the n differences : where is an observation and is the corresponding fitted value obtained.
Systematic Reviews Professor Kate O’Donnell. Reviews Reviews (or overviews) are a drawing together of material to make a case. These may, or may not,
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
The Effect of Computers on Student Writing: A Meta-Analysis of Studies from 1992 to 2002 Amie Goldberg, Michael Russell, & Abigail Cook Technology and.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Evidence Based Medicine Meta-analysis and systematic reviews Ross Lawrenson.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Biostatistics Case Studies 2007 Peter D. Christenson Biostatistician Session 3: Incomplete Data in Longitudinal Studies.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Funded through the ESRC’s Researcher Development Initiative Prof. Herb MarshMs. Alison O’MaraDr. Lars-Erik Malmberg Department of Education, University.
Meta-analysis and “statistical aggregation” Dave Thompson Dept. of Biostatistics and Epidemiology College of Public Health, OUHSC Learning to Practice.
The Campbell Collaborationwww.campbellcollaboration.org C2 Training: May 9 – 10, 2011 Introduction to meta-analysis.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Chapter 16 Data Analysis: Testing for Associations.
Developing a Review Protocol. 1. Title Registration 2. Protocol 3. Complete Review Components of the C2 Review Process.
Issues concerning the interpretation of statistical significance tests.
C82MST Statistical Methods 2 - Lecture 1 1 Overview of Course Lecturers Dr Peter Bibby Prof Eamonn Ferguson Course Part I - Anova and related methods (Semester.
Chapter 10 The t Test for Two Independent Samples
Correlation & Regression Analysis
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Systematic Synthesis of the Literature: Introduction to Meta-analysis Linda N. Meurer, MD, MPH Department of Family and Community Medicine.
Chi-Square Analyses.
Outline of Today’s Discussion 1.The Chi-Square Test of Independence 2.The Chi-Square Test of Goodness of Fit.
Handout Six: Sample Size, Effect Size, Power, and Assumptions of ANOVA EPSE 592 Experimental Designs and Analysis in Educational Research Instructor: Dr.
Funded through the ESRC’s Researcher Development Initiative Department of Education, University of Oxford Session 2.1 – Revision of Day 1.
Systematic Reviews and Meta-analyses. Introduction A systematic review (also called an overview) attempts to summarize the scientific evidence related.
URBDP 591 A Lecture 16: Research Validity and Replication Objectives Guidelines for Writing Final Paper Statistical Conclusion Validity Montecarlo Simulation/Randomization.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
An Application of Multilevel Modelling to Meta-Analysis, and Comparison with Traditional Approaches Alison O’Mara & Herb Marsh Department of Education,
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
Chapter 11: Test for Comparing Group Means: Part I.
Chapter 9 Introduction to the t Statistic
Chapter 10 Confidence Intervals for Proportions © 2010 Pearson Education 1.
Stats Methods at IC Lecture 3: Regression.
Systematic Reviews and Meta-analyses
Confidence Intervals.
DTC Quantitative Methods Bivariate Analysis: t-tests and Analysis of Variance (ANOVA) Thursday 20th February 2014  
RDI Meta-analysis workshop - Marsh, O'Mara, & Malmberg
Supplementary Table 1. PRISMA checklist
CJT 765: Structural Equation Modeling
Rehabilitation Research July 27, 2005
Meta-analysis April 11, 2006.
Presentation transcript:

Funded through the ESRC’s Researcher Development Initiative Department of Education, University of Oxford Session 2.3 – Publication bias

 A type of selection bias in which only published studies are analysed.  Assumption that studies that are published are more likely to report statistically significant findings.  Defined as “a bias against negative findings on the part of those involved in deciding whether to publish a study” (Soeken & Sripusanapan, 2003, p. 57).  It could lead to an overestimation of effect sizes, as published studies tend to have higher effect sizes than unpublished studies (Lipsey & Wilson, 1993).  Can result in inflation of Type I error rates in the published literature

 The debate about using only published studies:  peer-reviewed studies are presumably of a higher quality VERSUS  significant findings are more likely to be published than non-significant findings  There is no agreed upon solution. However, one should retrieve all studies that meet the eligibility criteria, and be explicit with how they dealt with publication bias.  Some methods for dealing with publication bias have been developed (e.g., Fail-safe N, Trim and Fill method).

 Inclusion of unpublished papers is likely to add considerable “noise” to the analyses  Methods typically used to find unpublished papers are ‘ad hoc’  the resulting selection of studies is likely to be less representative of the unknown population of studies than is the population of published studies, and typically will be more homogeneous (White, 1994).  “Whether bias is reduced or increased by including unpublished studies cannot formally be assessed as it is impossible to be certain that all unpublished studies have been located (Smith & Egger, 1998)”  Hence, for published papers, there is a more clearly defined population of studies to which to generalize than would be the case if unpublished studies were included.

 A central goal of meta-analysis is to be inclusive.  Meta-analyses call for a balance between practicality and comprehensiveness (Durlak & Lipsey, 1991).  A compromise is for meta-analysts to report how they dealt with publication bias

 Examination of the focus of the included studies  Fail-safe N  Trim & Fill  Sensitivity analysis (Vevea & Woods, 2005)

 Imagine a meta-analysis in which gender differences in self-concept is the focus  If the main aim of an included study is on a completely different issue, so that gender differences are incidental to the main focus of the study, the statistical significance of gender differences per se are unlikely to contribute to publication bias.  Hence, the generalisability of the effect size over this aspect of the study (degree to which gender effects are a main focus) provides a test of publication bias and an alternative perspective against which to validate other more generalisable approaches to this issue.

 Run an ANOVA in which effect sizes are grouped by 1. Studies in which the research question of the meta- analysis is central to the study (e.g., gender differences in self-concept) 2. Studies in which the research question of the meta- analysis is not central to the study (e.g., gender differences in achievement)  If significant difference in which studies with (1) as focus have higher mean effect size, then this could suggest publication bias (but does not confirm it!!!)

 The fail-safe N (Rosenthal, 1991) determines the number of studies with an effect size of zero needed to lower the observed effect size to a specified (criterion) level.  See Orwin (1983) for formula and discussion  For example, assume that you want to test the assumption that an effect size is at least.20.  If the observed effect size was.26 and the fail-safe N was found to be 44, this means that 44 unpublished studies with a mean effect size of zero would need to be included in the sample to reduce the observed effect size of.26 to.20.

 Trim and fill procedure (Duval & Tweedie, 2000a, 2000b) calculates the effect of potential data censoring (including publication bias) on the outcome of the meta-analyses.  Nonparametric, iterative technique examines the symmetry of effect sizes plotted by the inverse of the standard error. Ideally, the effect sizes should mirror on either side of the mean.

 The trim and fill process is as follows (O’Mara, 2008; see also Duval & Tweedie, 2000a, 2000b): 1. The overall mean is calculated in the usual manner. 2. The number of hypothetical missing studies is estimated and their corresponding non-missing effect sizes are “trimmed” from the dataset. 3. The overall mean is re-estimated (excluding the omitted effect sizes). 4. The newly calculated “trimmed” mean is used to re-estimate the number of hypothetical missing studies (as in Step One).

5. Steps Two to Four are repeated until the signed ranks of the effect sizes do not change, and the algorithm is terminated. 6. The original dataset is “filled” with symmetric data points representing the potential omitted studies. 7. The overall mean effect size and confidence intervals are recalculated, incorporating the influence of the hypothetical missing studies (Gilbody, Song, Eastwood & Sutton, 2000; Sutton, Duval, Tweedie, Abrams & Jones, 2000).

 The L 0 and R 0 estimators indicate how many hypothetical, non-published studies with a negative effect size were missing from the meta-analysis (step 6 on the previous slide)

 Assumption that all suppressed studies are those with the most negative effect sizes. If this assumption is inaccurate, the “corrected” effect- size estimate will be inaccurate (Vevea & Woods, 2005)

 The robustness of the findings to different assumptions can be examined through sensitivity analysis  It can be used to assess whether study quality affects the results, or whether publication bias is likely  Sensitivity analysis can take various forms. For example, can include study quality as a predictor to see if this has an impact on the effect sizes  In publication bias, Vevea & Woods (2005) applied it to a weight function

 “A new approach is proposed that is suitable for application to meta-analytic data sets that are too small for the application of existing methods. The model estimates parameters relevant to fixed- effects, mixed-effects or random-effects meta- analysis contingent on a hypothetical pattern of bias that is fixed independently of the data (p. 428)”  Imposes a set of fixed weights determined a priori and chosen to represent a specific form and severity of biased selection  Useful also when datasets are small

 “The author of the meta-analysis, then, is faced with a logically impossible task: to show that publication bias is not a problem for the particular data set at hand. We describe the task as logically impossible because it amounts, in essence, to an attempt at confirming a null hypothesis. (Vevea & Woods, 2005, p. 438)”  Different methods can attempt to address (assess?) the issue, but none is perfect.  At least we can conclude that the Fail-safe N is not appropriate!  Include unpublished studies?

 Duval, S., & Tweedie, R. (2000a). A Nonparametric "Trim and Fill" Method of Accounting for Publication Bias in Meta-Analysis. Journal of the American Statistical Association, 95,  Duval, S., & Tweedie, R. (2000b). Trim and fill: A simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis. Biometrics, 56, 455–463  Lipsey, M. W., & Wilson, D. B. (2001). Practical meta-analysis. Thousand Oaks, CA: Sage Publications.  Orwin, R. G. (1983). A Fail-Safe N for effect size in meta-analysis. Journal of Educational Statistics, 8,  Smith, G. D., & Egger, M. (1998). Meta-analysis: Unresolved issues and future developments. BMJ,  Van den Noortgate, W., & Onghena, P. (2003). Multilevel meta-analysis: A comparison with traditional meta-analytical procedures. Educational and Psychological Measurement, 63,  Vevea, J. L., & Woods, C. M. (2005). Publication bias in research synthesis: Sensitivity analysis using a priori weight functions. Psychological Methods, 10, 428–443.