A modest attempt at measuring and communicating about quality

Slides:



Advertisements
Similar presentations
WHERE TO NEXT? Using Reading Data. Group Learning Pathways.
Advertisements

Enhancing Data Quality of Distributive Trade Statistics Workshop for African countries on the Implementation of International Recommendations for Distributive.
Quality Guidelines for statistical processes using administrative data European Conference on Quality in Official Statistics Q2014 Giovanna Brancato, Francesco.
Screen 1 of 24 Reporting Food Security Information Understanding the User’s Information Needs At the end of this lesson you will be able to: define the.
OECD Short-Term Economic Statistics Working PartyJune Analysis of revisions for short-term economic statistics Richard McKenzie OECD OECD Short.
The use and convergence of quality assurance frameworks for international and supranational organisations compiling statistics The European Conference.
Chapter 9: Data quality and metadata Ilaria DiMatteo United Nations Statistics Division The 4 th meeting of the Oslo Group on energy statistics Ottawa,
Data Quality & dissemination D. Sahoo Dy. Director General Central Statistical Organization, India.
Stop the Madness: Use Quality Targets Laurie Reedman.
for statistics based on multiple sources
United Nations Oslo City Group on Energy Statistics OG7, Helsinki, Finland October 2012 ESCM Chapter 8: Data Quality and Meta Data 1.
Quality Frameworks: Implementation and Impact Notes by Michael Colledge.
Guidelines Recommandations. Role Ideal mediator for bridging between research findings and actual clinical practice Ideal tool for professionals, managers,
Copyright 2010, The World Bank Group. All Rights Reserved. QUALITY ASSURANCE AND EVALUATION Part 1: Quality Assurance 1.
Census quality evaluation: Considerations from an international perspective Bernard Baffour and Paolo Valente UNECE Statistical Division Joint UNECE/Eurostat.
Communication Arts The Writing Process. Communication Arts GUIDING CONCEPT As writers, we understand and demonstrate the ability and flexibility to use.
What is qualitative data analysis? Different approaches to analysing qualitative data.
Scientific Literature and Communication Unit 3- Investigative Biology b) Scientific literature and communication.
Quality declarations Study visit from Ukraine 19. March 2015
National Population Commission (NPopC)
Critical thinking for assignments to get a better grade
Comparative analysis (stage 2 English)
Information for Parents Key Stage 3 Statutory Assessment Arrangements
Presented by Deborah Eldridge, CAEP Consultant
Implementation of Quality indicators for administrative data
Theme (i): New and emerging methods
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 3. Session 9. Periodic data collection methods.
Development of Strategies for Census Data Dissemination
Tax Gap Analysis Presentation to GNPBO Conference June
Monitoring and Evaluation Systems for NARS Organisations in Papua New Guinea Day 2. Session 6. Developing indicators.
CMMI Q & A.
Programme Board 6th Meeting May 2017 Craig Larlee
Reading for Critical Thinking
Community program evaluation school
Writing Tasks and Prompts
OPCVL With reference to origin, purpose, and content, analyze the value and limitations.
"Development of Strategies for Census Data Dissemination".
How often do you get information from the Internet
Rigor is disciplinary…
4.1. Data Quality 1.
Internal Audit Strategy Survey Results & Discussion
Survey phases, survey errors and quality control system
Michael Biddington, UN ESCAP Statistics Division,
Measuring Data Quality and Compilation of Metadata
December 2016 Visitor Experience Introduction
Imagine your family is trying to decide whether or not to get a pet
Internet-based Research
Survey phases, survey errors and quality control system
Introduction. Conducting statistical investigations to develop learner statistical thinking.
6.1 Quality improvement Regional Course on
AESS Annual Assessment Process Workshop
Identification: Instrumental Variables
CORE Guaranteed & Viable Curriculum
Engaging with leaders Thursday 8th March 2011 Tim Heywood
CATHCA National Conference 2018
Sub-Regional Workshop on International Merchandise Trade Statistics Compilation and Export and Import Unit Value Indices 21 – 25 November Guam.
What to do with your data?
Quality evaluation of register-based statistics
Metadata used throughout statistics production
Zimbabwe 2008 Critical Thinking.
The Writing Process.
DESIGN OF EXPERIMENTS by R. C. Baker
Quality in statistics and relation to the media
Changing How We Communicate Quality and Methods to Our Users
Indicators, Data Sources and Data Quality for TB M&E
Kees Zeelenberg, Winfried Ypma, Peter Struijs; Statistics Netherlands
Metadata on quality of statistical information
Selecting Evidence to Demonstrate Teaching Effectiveness
Guy Van Gyes CAWIE-meeting 23-24/01/2012
READING AND RESPONDING
Presentation transcript:

A modest attempt at measuring and communicating about quality Laurie Reedman Statistics Canada Laurie.Reedman@Canada.ca June 28, 2018 Session 5

Outline Context Proposed quality indicator framework Metadata about quality Accuracy Next steps

Context Moving from survey-centric business model to survey- supported, or survey-free Making more use of data we did not produce Increasing appetite for data – more timely and more detail Quality challenge – how to estimate accuracy when we don’t have sampling error Research question at Statistics Canada: How should we measure and report quality in this context? We have just initiated a research project. We don’t have answers yet, we just have lots of questions.

Unpack the research question How to measure: Which dimensions or aspects of quality are important? How to report: Who is the audience? Stakeholders Internal External Casual data users Researchers Policy makers Other Let’s unpack the question. To figure out how to measure, we need to know what to measure, which dimensions or aspects of are important. To figure out how to report, we need to know the audience. I think we have several different audiences, and they all understand at different levels, And they need different information. Internally, we measure and monitor all aspects of quality, to manage our own resources. Externally, we have casual data users, such as journalists, who look at aggregate totals But don’t do complex analysis. Then we have power users, who could be researchers, or they could be policy and decision makers. There could very well be other users whose needs we do not yet understand. So perhaps we need to report quality in several different ways, in order to reach the different audiences.

Quality dimensions or aspects Users will judge for themselves Relevance Timeliness Interpretability Accessibility Coherence with other sources We measure (estimate) Accuracy Reliability Coverage Bias We assess when deciding to use data Perception of authority and credibility of data producer Quality Assurance practices followed by data producer Processability of the data Combinability or linkability of the data Coherence with standards When we talk about quality we break it down into different dimensions. The regular ones are still there but when we start using data that comes from different sources, There are additional dimensions or aspects of quality that we need to consider. This is not an exhaustive list. What I’ve tried to capture here is what are the key things to consider about quality from three different perspectives. At the input stage we evaluate these things and make a decision about whether we are going to use the data or not. At the throughout stage we do editing, data integration, analysis, imputation, calibration, and we measure or Estimate the accuracy, bias and coverage of the data. At the output stage we create metadata, and while we are honest and transparent about the data, We also want to showcase everything that is good about the data, so that people will trust it and use it. So for example users will judge for themselves if our data is relevant for their purposes, but we can inform Their decision by telling them where it comes from, what the concepts, what is the reference period, What is the coverage.

DRAFT Quality Indicator Framework Monitor Internally Report Externally Quality Dimension or Aspect Yes Credibility QA practices followed Maybe not Processability Linkability Coherence with standards Accuracy Maybe Reliability Coverage Bias Sort of Relevance Timeliness Metadata Interpretability Accessibility yes Coherence with other sources Quality Indicator Framework DRAFT It was deliberate that I wrote DRAFT across this slide and made it difficult for you to read. Remember that we have just started our research project at Statistics Canada, and we don’t have answers yet. This is an early draft of a quality indicator framework. In the right-hand column it has all of the quality dimensions from the previous slide. On the left is a whole column of Yes’s showing that we measure and monitor all of these things internally. The fun is in the middle column. Here I am trying to show which dimensions or aspects we should report to data users. When I showed this slide to my colleagues we had a big discussion about which ones should be yes And which ones should be maybe or maybe not. We don’t agree yet which ones we should report, never mind how to report them. That’s why it’s a research project. Let’s look at linkability for example, that one we agreed is a “yes”. When we are deciding whether to use a particular dataset we consider how easy It will be to link it with other data sources that we have. We could report to data users the linkage rate. We might also want to point out to users what variables are on the dataset that they can use to do their own linkages. This would help them assess the linkability or usefulness of our dataset for their purposes.

Metadata about quality Quality indicator framework List, like a nutritional label Composite index Infographic (spiderweb, weather symbols, …) Assumptions, compromises and limitations Why we made assumptions and compromises, what is the impact on quality and usability of the data, what are the limitations of the data Certification What is the standard, who did the assessment, what’s included (Only the product? Also the process?) I like the idea of a list like on a nutritional label, where all the elements are in the same order And are reported with standardized units of measure. A composite quality index would be really cool but I think we are a long way from being able to build one. Infographics convey the least amount of information in the quickest amount of time. These might be Ideal for the casual user but would likely not satisfy all our users. So maybe we need different layers to the metadata about quality that target different audiences. For a long time now we have been explaining in our metadata the assumptions and compromises we Make, but maybe we should say a bit more about the impact these things have on the quality or usability of the data. What limitations are there, what we would recommend you not try to do with this data. If we are going to have some sort of certification then we should describe what the standards are, and ideally it would Be an independent party doing the assessment, so we should say who that is. I would also like to see the assessment look at the quality of statistical processes as well as the quality of statistical products.

A few words about accuracy How to estimate accuracy and reliability when there is no sampling error Accuracy – how close are the numbers to reality Reliability – how accurate are the numbers through time Non-sampling errors can be: Systematic – resulting in bias Random – resulting in increased variability (noise) To estimate bias – data validation, confrontation, resampling methods (Statistics Netherlands report, 2014) To estimate noise – the bootstrap method has potential under certain circumstances

Next steps Research project at Statistics Canada Don’t want to re-invent the wheel We welcome collaboration on methods, standards, terminology

A modest attempt at measuring and communicating about quality Thank you / Merci Laurie Reedman, Statistics Canada