The Use of Paradata to Improve Survey Quality: Organizational Approaches and Challenges March 27, 2012 Ananth Koppikar.

Slides:



Advertisements
Similar presentations
Identify real savings, collaborate more effectively, improve contract compliance and evidence your success.
Advertisements

C6 Databases.
“... providing timely, accurate, and useful statistics in service to U.S. agriculture.” Using Mixed Methods to Evaluate Survey Questionnaires Heather Ridolfo,
Enhancing Data Quality of Distributive Trade Statistics Workshop for African countries on the Implementation of International Recommendations for Distributive.
May 16, 2014 Analytics Short Discussion. ECAR Analytics Maturity Index Source: ECAR Analytics Maturity Index, 2012.
2011 OSEP Leadership Mega Conference Collaboration to Achieve Success from Cradle to Career 2.0 State Monitoring Under IDEA A Snapshot of Past Practices.
Survey Design Steps in Conducting a survey.  There are two basic steps for conducting a survey  Design and Planning  Data Collection.
Customer Relationship Management..a strategy used to learn more about customers' needs and behaviours in order to develop stronger relationships with them.
Army Cost Management & Financial Transparency
Why Managers Must Understand IT Managers play a key role –Frame opportunities and threats so others can understand them –Evaluate and prioritize problems.
Business Intelligence Dr. Mahdi Esmaeili 1. Technical Infrastructure Evaluation Hardware Network Middleware Database Management Systems Tools and Standards.
Enable Business Intelligence Across the Enterprise William McKnight McKnight Associates, Inc.
Data Warehouse Components
08/08/2015 Statistics Canada Statistique Canada Paradata Collection Research for Social Surveys at Statistics Canada François Laflamme International Total.
18/08/2015 Statistics Canada Statistique Canada Responsive Collection Design (RCD) for CATI Surveys and Total Survey Error (TSE) François Laflamme International.
Microsoft Office Project Portfolio Server
By N.Gopinath AP/CSE. Why a Data Warehouse Application – Business Perspectives  There are several reasons why organizations consider Data Warehousing.
Using survey data collection as a tool for improving the survey process Silvia Biffignandi, Antonio Laureti Giulio Perani University of Bergamo Istat Istat.
1.Knowledge management 2.Online analytical processing 3. 4.Supply chain management 5.Data mining Which of the following is not a major application.
CIS 429—Chapter 8 Accessing Organizational Information—Data Warehouse.
Best Practices: Aligning Process, Culture and Tools Michael Jordan Senior Project Manager - Microsoft Consulting Services
Data Profiling
Too expensive Too complicated Too time consuming.
GBA IT Project Management Final Project - Establishment of a Project Management Management Office 10 July, 2003.
Data Warehouse Overview September 28, 2012 presented by Terry Bilskie.
ETL Overview February 24, DS User Group - ETL - February ETL Overview “ETL is the heart and soul of business intelligence (BI).” -- TDWI ETL.
OECD/INFE Tools for evaluating financial education programmes Adele Atkinson, PhD Policy Analyst OECD With the support of the Russian/World Bank/OECD Trust.
Measuring Respondent Burden at the UK Office for National Statistics Denise Williams, Sarah Green & Jacqui Jones.
111 Notion of a Project Notes from OOSE Slides – a different textbook used in the past Read/review carefully and understand.
Internet Surveys & Future Works In Official Statistics 1 The 4th International Workshop on Internet Survey Methods, Sept. Statistical Center, Statistics.
CZECH STATISTICAL OFFICE 1 The Quality Metadata System In the Czech Statistical Office Work Session on Statistical Metadata (METIS)
Research Designs & Methods  Research design: the strategy and overall framework of the research. Example: A survey design.  Research method: how the.
1 Renee M. Gindi NCHS Federal Conference on Statistical Methodology Statistical Policy Seminar December 4, 2012 Responsive Design on the National Health.
Data Warehouse Development Methodology
2 Copyright © Oracle Corporation, All rights reserved. Defining Data Warehouse Concepts and Terminology.
Soup-2-Nuts Alaska Department of Fish & Game Commercial Fisheries October, 2011.
C6 Databases. 2 Traditional file environment Data Redundancy and Inconsistency: –Data redundancy: The presence of duplicate data in multiple data files.
Quality Assurance Programme of the Canadian Census of Population Expert Group Meeting on Population and Housing Censuses Geneva July 7-9, 2010.
Statistics Canada Statistique Canada Cost-Efficient Framework for Data Collection for CATI Surveys Social Surveys Collection Research Steering Committee.
Statistical Expertise for Sound Decision Making Quality Assurance for Census Data Processing Jean-Michel Durr 28/1/20111Fourth meeting of the TCG - Lubjana.
Question paper 1997.
Building Dashboards SharePoint and Business Intelligence.
©2005 Prentice Hall Business Publishing, Introduction to Management Accounting 13/e, Horngren/Sundem/Stratton ©2008 Prentice Hall Business Publishing,
Consultant Advance Research Team. Outline UNDERSTANDING M&E DATA NEEDS PEOPLE, PARTNERSHIP AND PLANNING 1.Organizational structures with HIV M&E functions.
Metadata By N.Gopinath AP/CSE Metadata and it’s role in the lifecycle. The collection, maintenance, and deployment of metadata Metadata and tool integration.
Arlington, VA March 31, 2004 Presentation for the Advisory Committee for Business & Operations Effective Practices Research Overview For Merit Review This.
Sampling Design and Analysis MTH 494 Ossam Chohan Assistant Professor CIIT Abbottabad.
Onsite Quarterly Meeting SIPP PIPs June 13, 2012 Presenter: Christy Hormann, LMSW, CPHQ Project Leader-PIP Team.
Monitoring Afghanistan, 2015 Food Security and Agriculture Working Group – 9 December 2015.
Introductory Session: Welcome to the Universal Standards Implementation Series Speaker: Yamini Annadanam 16 October 2014.
1 Copyright © 2009, Oracle. All rights reserved. Oracle Business Intelligence Enterprise Edition: Overview.
9 - 1 Chapter 9 Management Control Systems and Responsibility Accounting.
Chapter 8: Data Warehousing. Data Warehouse Defined A physical repository where relational data are specially organized to provide enterprise- wide, cleansed.
The Concepts of Business Intelligence Microsoft® Business Intelligence Solutions.
Overview of Types of Measures Margaret Kasimatis, PhD VP for Academic Planning & Effectiveness.
2 Copyright © 2006, Oracle. All rights reserved. Defining Data Warehouse Concepts and Terminology.
Shared Services Initiative Summary of Findings and Next Steps.
Best Practices for Implementing a Paradata Warehouse Presentation at the Washington Statistical Society Mini-Conference on Paradata Washington, D.C. Jason.
Division of HIV/AIDS Managing Questionnaire Development for a National HIV Surveillance Survey, Medical Monitoring Project Jennifer L Fagan, Health Scientist/Interview.
JMFIP Financial Management Conference
Building a Data Warehouse
Database Development (8 May 2017).
Scientific Information Management Approaches Needed to Support Global Assessments: The USEPA's Experience and Information Resources Jeffrey B. Frithsen.
Automating Security Frameworks
Data Warehouse.
Survey phases, survey errors and quality control system
Multi-Mode Data Collection Approach
Data Warehouse Overview September 28, 2012 presented by Terry Bilskie
Survey phases, survey errors and quality control system
Database Development Cycle
Presentation transcript:

The Use of Paradata to Improve Survey Quality: Organizational Approaches and Challenges March 27, 2012 Ananth Koppikar

I.What paradata are collected? II.Are standardized paradata being collected across modes and/or studies? III.Is there a set of key indicators used to assess survey process quality? IV.Conceptual paradata system architecture at Mathematica V.Operational challenges VI.Costs and justification Overview 2

 Paradata indicators currently collected and reported on include: cost per complete, outbound calls per hour, clerical support, survey operations support, supervision hours per complete, and monitoring and mileage per complete (as appropriate by mode).  Paradata indicators have been classified in three broad functional categories of interest: (1) production and cost, (2) data quality, and (3) business development.  All indicators are associated with specific operational questions, such as: Are Survey Operations Center (SOC) and supervision hours per complete in line with budget expectations? I.What paradata are collected? 3

 Paradata targeted for the future phases include: –Date and time an interview was started (each CAI mode) –Date and time an interview was completed (each CAI mode) –Questionnaire responses –Average time per complete or average number of attempts –Item-level Don’t Knows/Refused/Blanks –Key strokes – changing answers or backing up –Contact data Date/time Time slice Contact result codes/response rates –Number of interviewers and number of hours worked  In addition, we may capture and study the correlation of some paradata indicators to certain project metadata such as characteristics of interviewer and respondent, incentives, calling rules, etc. I.What paradata are collected? (continued) 4

 The paradata that we have identified for inclusion in 2012 are captured in standardized formats across projects.  Standardized call details and history are available on our Blaise CATI system.  For other modes, most projects use our Survey Management System (SMS), and its underlying database structure permits standardization of sample and operational data across projects.  Other standard data sources include Deltek’s Time & Expense system and GovWin.  While standard elements are being captured in key systems, important challenges remain in developing standardized usage/indicator values. II.Are standardized paradata being collected across modes and/or studies? 5

 Key indicators by mode: –Cost per complete (TINT, Field, Web) –Clerical hours per complete (TINT, Field) –Hours per complete (TINT, Web) –Mileage per complete (Field) –Monitoring percentages (completes, cases per interviewer, TINT hours) –Outbound calls per hour (TINT) –SOC support hours (TINT, Field) –Supervision percentage (TINT)  These paradata shed light on the following operational questions: –Are costs per complete higher or lower than expected? –Are support, on-call, clerical, monitoring, and supervision hours in line with the project budget? –Are hours per complete higher or lower than expected? –Is call volume higher or lower than expected? –Is mileage per complete higher or lower than expected? –Is outbound call volume higher or lower than expected? III.Is there a set of key indicators used to assess survey process quality? 6

CAI CostPoint GovWin SMS Other Sources of Paradata and Metadata Data Warehouse (all projects, rounds, etc.) Selected Paradata and Metadata Reports, Charts, Dashboards, etc. Business Intelligence Tools, Reporting Tools, etc. IV. Conceptual Paradata System Architecture at Mathematica Extract, Transform, and Load 7

 To collecting and processing paradata: –Designing a flexible, scalable data warehouse that can store selected paradata –Standardizing values across projects –Missing data  To collecting standardized paradata and producing standard indicators across modes and/or studies: –There is a cultural and educational challenge to the adoption of standards Overcoming the fear that standardization leads to loss of important detail and that project-specific information needs will go unmet Overcoming the fear that standardization will be too costly and will not provide a good return on investment Once the case is made for standards, the challenge remains of which standards to adopt  We are investigating adoption of emerging paradata standards V.Operational Challenges 8

 To providing tailored indicators to address study-specific design issues: –Existing paradata reports are not widely or consistently used –Development of new report templates that are readily interpretable and contain actionable data has proven difficult  To Implementing responsive designs based on process quality indicators –A model for implementing responsive survey design is being developed as part of the paradata initiative. –This model is based on the Total Survey Error (TSE) approach, which assumes a tradeoff between survey error and survey cost and provides a framework for reducing respondent selection error, response accuracy issues, and survey administration errors, as well as managing cost. –Achieving consensus on an organization-wide model is a time-consuming process –Training staff to understand paradata and use it appropriately V.Operational Challenges (continued) 9

 Costs associated with developing a paradata system will vary based on the complexity of the system.  Costs include labor associated with researcher and programmer time, hardware (e.g., servers), and software (e.g., BI tools).  Initial phase at Mathematica (developing overall framework; formulating research questions; drilling down to identify indicators; specifying requirements for indicators and reports; designing and building the data warehouse, ETL programs, and reports; piloting and then deploying to production) is expected to cost over $200K. For this initial phase, we will be using existing hardware and not incorporating any BI tools. VI.Costs and Justification 10

 Regular, readily interpretable reporting on paradata provides the foundation for the implementation of responsive survey design  Anecdotal evidence exists at Mathematica that paradata have been used on some projects to reduce data collection costs and improve response rates.  Senior management recognizes that paradata would be an invaluable tool to improving data quality and optimizing costs, among other things.  Costs are viewed by senior management as an investment and not an expense.  Approaching the paradata initiative in phases – helps manage costs and expectations, and provides senior management with a greater degree of control on the investment. VI.Costs and Justification 11