Nicole R. Buttermore, Frances M. Barlas, & Randall K. Thomas

Slides:



Advertisements
Similar presentations
Method Introduction Results Discussion The Effect of Self-Esteem, Marital Status, and Gender on Trait Anxiety and Stress Emily B Gale University of Nebraska-Lincoln.
Advertisements

The Effect of Predisposing Factors and Concussion Rate on DIII College Football Players: A Retrospective Study Jon Purvis, Robert Blume, Jenna Chinburg,
A Deeper Understanding of Avery Fitness Center Customers
Populations and Samples Anthony Sealey University of Toronto This material is distributed under an Attribution-NonCommercial-ShareAlike 3.0 Unported Creative.
Advanced Business Research Method Intructor : Prof. Feng-Hui Huang Agung D. Buchdadi DA21G201.
 Closing the loop: Providing test developers with performance level descriptors so standard setters can do their job Amanda A. Wolkowitz Alpine Testing.
Experimental Design If a process is in statistical control but has poor capability it will often be necessary to reduce variability. Experimental design.
BUSINESS STATISTICS MGT 2302 BUSINESS STATISTICS MGT 2302 Lecturer Name : Liyana ‘Adilla 1 SCHOOLOGY ACCESS CODE: 7QRB9-4MPNN.
Buying and Selling Prices under Risk, Ambiguity and Conflict Michael Smithson The Australian National University Paul D. Campbell Australian Bureau of.
Family Program Needs Select results from the 2008 Active Duty Spouse Survey (ADSS2008) and Reserve Component Spouse Survey (RCSS2008) January 2011.
Researching Technology in South Dakota Classrooms Dr. Debra Schwietert TIE Presentation April 2010 Research Findings.
1 Collecting and Interpreting Quantitative Data Deborah K. van Alphen and Robert W. Lingard California State University, Northridge.
Parental, Temperament, & Peer Influences on Disordered Eating Symptoms Kaija M. Muhich, Alyssa Collura, Jessica Hick and Jennifer J. Muehlenkamp Psychology.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
Multivariate Analysis - Introduction. What is Multivariate Analysis? The expression multivariate analysis is used to describe analyses of data that have.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 1-1 Statistics for Managers Using Microsoft ® Excel 4 th Edition Chapter.
WHO The World Health Survey HOUSEHOLD QUESTIONNAIRE
Candace Pang1 and Elizabeth Price2; Mentors: Dr. Chien-fei Chen3, Dr
Survey Methodology Reliability and Validity
Learning Objectives : After completing this lesson, you should be able to: Describe key data collection methods Know key definitions: Population vs. Sample.
Contribution of Physical Education to Physical Activity of Children
Better to Give or to Receive?: The Role of Dispositional Gratitude
Quantified perceived and Expected Inflation in the Euro Area
Survey Design with the Office of Assessment
National Consumer Agency
Prevalence of Adverse Childhood Experiences
Quality of Electronic Emergency Department Data: How Good Are They?
Emergency department pediatric psychiatric services
Media Comparisons 2016.
A Comparison of Two Nonprobability Samples with Probability Samples
Michigan Future Business Index
Assessing Students' Understanding of the Scientific Process Amy Marion, Department of Biology, New Mexico State University Abstract The primary goal of.
The Relationship Between Emphasis of Cell-phone Use on Performance and Anxiety: Classroom Implications Jordan Booth, Leah Cotton, Jeni Dillman, Kealey.
Multivariate Analysis - Introduction
SHRM Survey Findings: State of Employee Benefits in the Workplace—Communicating Benefits January 10, 2013.
Assessments Attitudes of UK Teachers & Parents
Take-home quiz due! Get out materials for notes!
US Consumer Analysis: Apparel and Footwear February, 2017.
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
DESCRIPTIVES AND CORRELATIONS
Chapter 4 PRESENTATION AND ANALYSIS OF DATA
Research amongst Physical Therapists in the State of Kuwait: Participation, Perception, Attitude and Barriers Presented by Sameera Aljadi, PT, PhD Assistant.
My, But We are Impressive
Chapter 7 The Hierarchy of Evidence
Business and Management Research
Chapter Three Research Design.
Making Causal Inferences and Ruling out Rival Explanations
Analyzing Reliability and Validity in Outcomes Assessment Part 1
Lecture 2: Data Collecting and Sampling
Asist. Prof. Dr. Duygu FIRAT Asist. Prof.. Dr. Şenol HACIEFENDİOĞLU
Cross Sectional Designs
Determinants of Household Allocation of Income in Iceland
Research Methods Week 4 Research Design.
Chapter 1 The Where, Why, and How of Data Collection
Chapter 1 The Where, Why, and How of Data Collection
Business and Management Research
11th Annual Parents, kids & money survey
Analyzing Reliability and Validity in Outcomes Assessment
Introduction to the design (and analysis) of experiments
Measuring Outcomes and Designing Research
Multivariate Analysis - Introduction
The Where, Why, and How of Data Collection
NATIONAL ASSOCIATION OF REALTORS®
Levine et al continued.
Collecting and Interpreting Quantitative Data
CLASSROOM ENVIRONMENT AND THE STRATIFICATION OF SENIOR HIGH SCHOOL STUDENT’S MATHEMATICS ABILITY PERCEPTIONS Nelda a. nacion 5th international scholars’
Tests are given for 4 primary reasons.
Chapter 13 Excel Extension: Now You Try!
NHS DUDLEY CCG Latest survey results August 2018 publication.
Chapter 1 The Where, Why, and How of Data Collection
Presentation transcript:

Nicole R. Buttermore, Frances M. Barlas, & Randall K. Thomas Consistency is Key: Replication across Online Non-probability Sample Sources Nicole R. Buttermore, Frances M. Barlas, & Randall K. Thomas GfK Custom Research

Introduction With the growing interest in online non-probability samples, one emerging concern in survey science is the ability to replicate effects across different sample sources. Research Question: Does what you find with one non-probability sample source replicate with other sample sources?

Introduction In the social sciences, a number of studies have had difficulties replicating both the direction and degree of experimental effects across convenience samples (Nozek et al, 2015; Heinrich, Heine, & Norenzayan, 2010). However, other researchers have found that experimental results obtained in convenience, non-probability samples are roughly equivalent in their accuracy across samples (Yeager et al., 2011).

Introduction Our interest in this study was to examine experimental replications across a wide range of non-probability sample providers. The Foundations of Quality 2 (FoQ2) was a major study investigating survey quality conducted in 2013 by the Advertising Research Foundation. The study collected data from 17 different non-probability sample providers with each providing three separate samples averaging about 1,100 respondents.

Introduction The FoQ2 dataset has been used extensively for other purposes: Research with the FoQ2 dataset compared results across non-probability sample providers for 24 national benchmarks (Gittelman, Thomas, Lavrakas, & Lange, 2015). The study found significant variance across providers in their average proximity to the benchmarks. Further, the use of respondent quotas and use of demographic weighting did not consistently reduce the divergences.

Method

Method The ARF FOQ2 dataset had 57,104 completes from 17 different opt-in sample sources (with 3 samples each = 51 different samples). The study fielded in January 2013. The study’s primary intent was to examine how using different sample quota methods might impact the quality of research results.

Method In this study, there were four major content areas we could examine for replicability of findings included: An experiment on response formats for past six month purchase, An experiment on 4 different new product concepts, Evaluations of 27 different brands, and Rated order of political spending priorities.

Method To assess replicability, there were two major measures for each content area: Average absolute divergence from the overall mean across samples Relative ordering of results (i.e., high ratings for the same brand across samples compared to low ratings for another brand). We used demographically weighted data for comparisons to eliminate confounding factors due to demographic differences across sample providers.

Results

Test 1 Summary Respondents were asked about their past 6 month purchase of a series of 6 items (randomly selected from 10 items). Men’s clothing Women’s clothing Children’s clothing Home improvement items such as tools or paint Automotive accessories or parts Groceries Sporting goods Computer-related electronics Entertainment-related products (DVDs, stereos, TVs, etc.) A Segway transporter

Test 1 Summary Respondents were randomly assigned to use one of three response formats: Multiple response Yes-no grid Combination grid Based on prior research, we have found that multiple response format (‘Select all’) yields significantly lower endorsement levels than yes-no grids, and yes-no grids have lower endorsement levels than combination grids.

Purchase Response Format The experimental results were replicated in 50 of the 51 samples, and partially replicated in the 1 remaining sample. All showed the highest endorsement frequency in the combination grid.

Purchase Response Format Average mean difference of endorsement between response format conditions. Difference between means

Test 2 Summary Respondents were randomly assigned 1 out of 4 possible new product concepts, they read a description of the new concept, and then rated their likelihood to purchase the product. The 4 concepts were: OneSurface Paint Soup-to-Go Energy-on-Demand Unbeatable Luggage

New Product Preferences The experimental results were replicated in 45 of the 51 samples, and partially replicated in the remaining 6 samples. The concept with highest purchase interest was ‘OneSurface Paint’ with 51 rating it highest; 47 samples rated ‘Energy-on-Demand’ lowest.

Test 3 Summary Respondents were randomly assigned 5 out of 27 brands to evaluate how much they liked each brand (with a ‘Not familiar’ response available). Analyses excluded ‘Not familiar’ responses.

Brand Ratings The ratings of brands remained fairly consistent across samples, with most samples having a correlation with the overall brand ratings (pooling total sample) of over .95.

Brand Ratings The overall ordering of brands was fairly consistent across samples. 0 = Strongly Dislike; 100 = Strongly Like

Test 4 Summary Respondents were presented with 6 programs that faced budget cuts in the U.S. and were asked which 1 they would reduce spending on the most. 1 Military weapons programs 2 Border patrol and prevention of illegal immigration 3 Police and crime prevention 4 Development of new technology 5 Health care system Social welfare system for the poor

Political Priorities The overall results were replicated in 28 of the 51 samples, but the top and bottom programs selected for a spending cuts were replicated in all 51 samples. The second and third top programs for program cuts alternated in a number of samples, whereas the fourth and fifth programs were more reliably replicated.

Conclusions and Discussion

Summary of Results For the purchase response format experiment and new product evaluation experiment, the relative rates of endorsement and new product preferences remained fairly constant. For the brand ratings, the overall ordering of brands was fairly consistent across samples, and we saw less absolute divergence among the highest rated brands. Finally, for political priorities, the experimental results were replicated in only 28 of 51 samples, primarily due to the proximity of the second and third priorities for spending cuts, which crossed over for a number of the samples.

Conclusions and Discussion Overall, we found most findings replicated across non-probability samples. One issue is not just the relative differences among samples, but also the absolute values – this is something we did not study. Another issue we are studying further – a subset of samples of the 51 were from river samples. Among these samples, replication of results seemed somewhat more problematic.

Thank you! Contact: Nicole R. Buttermore email: nicole.buttermore@gfk.com