Q2010, Helsinki Development and implementation of quality and performance indicators for frame creation and imputation Kornélia Mag László Kajdi Q2010,

Slides:



Advertisements
Similar presentations
Improving imputation methodology in the Hungarian Central Statistical Office (HCSO) NTTS 2009 seminar, Bruxelles February 2009 Improving imputation.
Advertisements

Module B-4: Processing ICT survey data TRAINING COURSE ON THE PRODUCTION OF STATISTICS ON THE INFORMATION ECONOMY Module B-4 Processing ICT Survey data.
Enhancing Data Quality of Distributive Trade Statistics Workshop for African countries on the Implementation of International Recommendations for Distributive.
Quality assurance -Population and Housing Census Alma Kondi, INSTAT, Albania.
Quality Guidelines for statistical processes using administrative data European Conference on Quality in Official Statistics Q2014 Giovanna Brancato, Francesco.
1 Editing Administrative Data and Combined Data Sources Introduction.
MISUNDERSTOOD AND MISUSED
Documentation and survey quality. Introduction.
Chapter Three Research Design.
United Nations Economic Commission for Europe Statistical Division Applying the GSBPM to Business Register Management Steven Vale UNECE
Environment Change Information Request Change Definition has subtype of Business Case based upon ConceptPopulation Gives context for Statistical Program.
Arun Srivastava. Types of Non-sampling Errors Specification errors, Coverage errors, Measurement or response errors, Non-response errors and Processing.
Eurostat Repeated surveys. Presented by Eva Elvers Statistics Sweden.
European Conference on Quality in Official Statistics (Q2010) 4-6 May 2010, Helsinki, Finland Brancato G., Carbini R., Murgia M., Simeoni G. Istat, Italian.
Use of survey (LFS) to evaluate the quality of census final data Expert Group Meeting on Censuses Using Registers Geneva, May 2012 Jari Nieminen.
List frames area frames and administrative data, are they complementary or in competition? Elisabetta Carfagna University of Bologna Department of Statistics.
[Hayes, Dassen, Schilder and Wallage, Principles of Auditing An Introduction to ISAs, edition 2.1] © Pearson Education Limited 2007 Slide 8.1 Control Risk,
Quality issues on the way from survey to administrative data: the case of SBS statistics of microenterprises in Slovakia Andrej Vallo, Andrea Bielakova.
Assessing Quality for Integration Based Data M. Denk, W. Grossmann Institute for Scientific Computing.
Eurostat Overall design. Presented by Eva Elvers Statistics Sweden.
Quality framework for the evaluation of administrative data (to be used for statistics) Piet J.H. Daas, Judit Arends-Tóth, Barry Schouten and Léander Kuivenhoven.
European Conference on Quality in Official Statistics Session 26: Quality Issues in Census « Rome, 10 July 2008 « Quality Assurance and Control Programme.
for statistics based on multiple sources
United Nations Economic Commission for Europe Statistical Division Mapping Data Production Processes to the GSBPM Steven Vale UNECE
Use of Administrative Data Seminar on Developing a Programme on Integrated Statistics in support of the Implementation of the SNA for CARICOM countries.
DATA PREPARATION: PROCESSING & MANAGEMENT Lu Ann Aday, Ph.D. The University of Texas School of Public Health.
Quality Assurance Programme of the Canadian Census of Population Expert Group Meeting on Population and Housing Censuses Geneva July 7-9, 2010.
© Federal Statistical Office Germany, Division IB, Institute for Research and Development in Federal Statistics Sheet 1 Surveys, administrative data or.
1 C. ARRIBAS, D. LORCA, A. SALINERO & A. COLMENERO Measuring statistical quality at the Spanish National Statistical Institute.
The challenge of a mixed-mode design survey and new IT tools application: the case of the Italian Structure Earning Surveys Fabiana Rocci Stefania Cardinleschi.
Outlining a Process Model for Editing With Quality Indicators Pauli Ollila (part 1) Outi Ahti-Miettinen (part 2) Statistics Finland.
Pilot Census in Poland Some Quality Aspects Geneva, 7-9 July 2010 Janusz Dygaszewicz Central Statistical Office POLAND.
Evaluating the Quality of Editing and Imputation: the Simulation Approach M. Di Zio, U. Guarnera, O. Luzi, A. Manzari ISTAT – Italian Statistical Institute.
United Nations Oslo City Group on Energy Statistics OG7, Helsinki, Finland October 2012 ESCM Chapter 8: Data Quality and Meta Data 1.
3-1 Copyright © 2010 Pearson Education, Inc. Chapter Three Research Design.
Copyright 2010, The World Bank Group. All Rights Reserved. Recommended Tabulations and Dissemination Section B.
Sampling Design and Analysis MTH 494 Ossam Chohan Assistant Professor CIIT Abbottabad.
QUALITY ASSESSMENT OF THE REGISTER-BASED SLOVENIAN CENSUS 2011 Rudi Seljak, Apolonija Flander Oblak Statistical Office of the Republic of Slovenia.
Towards a Process Oriented View on Statistical Data Quality Michaela Denk, Wilfried Grossmann.
Census quality evaluation: Considerations from an international perspective Bernard Baffour and Paolo Valente UNECE Statistical Division Joint UNECE/Eurostat.
5.8 Finalise data files 5.6 Calculate weights Price index for legal services Quality Management / Metadata Management Specify Needs Design Build CollectProcessAnalyse.
14-Sept-11 The EGR version 2: an improved way of sharing information on multinational enterprise groups.
The business process models and quality issues at the Hungarian Central Statistical Office (HCSO) Mr. Csaba Ábry, HCSO, Methodological Department Geneva,
Quality at a Glance: Documentation of Quality Indicators at Statistics Austria European Conference on Quality in Official Statistics Rome, 8-11 July 2008.
FDI - Imputation. Overview Introduction Overview of Imputation Methods Overview of Outliering methods Overview of Estimation methods Aggregation Disclosure.
How to deal with quality aspects in estimating national results Annalisa Pallotti Short Term Expert Asa 3st Joint Workshop on Pesticides Indicators Valletta.
Methods for Data-Integration
National Population Commission (NPopC)
Implementation of Quality indicators for administrative data
Analysis of the Reporting Forms for the Structural Business Survey
Session 7.1 Data Quality 1.
Chapter Three Research Design.
4.1. Data Quality 1.
An Active Collection using Intermediate Estimates to Manage Follow-Up of Non-Response and Measurement Errors Jeannine Claveau, Serge Godbout and Claude.
Prague EU-SILC Best Practice Workshop, 14th and 15th September 2017
Survey phases, survey errors and quality control system
KOMUSO Information for the Big Data society in official statistics
Survey phases, survey errors and quality control system
VAT data in Business Register and Business Statistics
6.1 Quality improvement Regional Course on
Issues in Administrative Data
Measurement errors Marina Signore
Mapping Data Production Processes to the GSBPM
Quality Reporting in CBS
The Swedish survey on turnover in the service sector
The role of metadata in census data dissemination
2.7 Annex 3 – Quality reports
SILC draft implementing and delegated acts Item 3.4 of the agenda
Presentation transcript:

Q2010, Helsinki Development and implementation of quality and performance indicators for frame creation and imputation Kornélia Mag László Kajdi Q2010, Helsinki

Introduction, background  Eurostat financed grant (planned finish: August 2010.)  The final report of the project will contain the following results regarding to frame creation and imputation:  Identification of the inputs, the result and the impacts of process steps and the flow charts.  Identification of the dimensions of process quality  Detailed description of the specified standard variables like required metadata, calculation and the possible variations of the variables, as well as recommendations for their application. The possible implementation of the indicators and the assessment of the measurement capability will be presented in the case of a concrete survey.  The development is based on the methods discussed in „Handbook on improving quality by analysis of process variables” (Eurostat [2003])”

Q2010, Helsinki International experiences Most often used and recommended indicators:  For frame selection: coverage errors (over- and undercoverage, misclassification and duplication)  For imputation: imputation rate and ratio  International experiences  Eurostat Standard Quality Indicators  ONS indicators  Hungarian experiences Process Quality Strategy Project:  Process variables for the production process steps

Q2010, Helsinki Frame creation The fact that we can not observe the whole target population causes different errors on survey estimations. Quality components  Relevance: any differences in the definition of the target population and the frame population or in the auxiliary information which are used in the survey can generate deficiencies on the relevance of the estimation.  Accuracy: any coverage problem that arises from not being able to sample from the whole of the target population can lead to the bias of the final estimation.  Timelines of the frames: the time lag between the reference date of the frame and the survey can cause coverage problems and consequently, it can have an effect on the final accuracy.

Q2010, Helsinki Frame creation – process flow Survey information:  Target population definition  Auxiliary information  Reference period Frame description:  Unit definition  Auxiliary information definition  Reference periods of the units/items Asessment procedures Differences of the unit definitions Differences of the auxiliary information definitions Time lag between the reference periods Deviation in the data Errors Missing units Non population elements Multiplicity elements Incorrect or missing auxiliary information

Q2010, Helsinki Process indicators  Overcoverage I.: Number and percentage of exisiting units in the frame that are not elements of the target population  Overcoverage II.: Number and percentage of non-existing units in the frame  Undercoverage.: Number and percentage of units which are not in the frame, but are part of the second frame  Misclassification: Number and percentage of units that are misclassified  Number of multiplied units  Number and percentage of the units excluded because of the time lag between reference dates of the survey and the frame Calculation: Number of units in the frame constructed at a later stage, but excluded from the original frame / (Number of units in the frame used for the survey) * 100 Metadata: Elements of the frame used for the survey; elements of the frame constructed at a later stage Comment: This can be calculated if we use the quality check approach, which means portions of the frame are independently constructed at a time close to the survey.  Number and percentage of units with missing auxiliary information (by type)

Q2010, Helsinki Imputation Imputation is the method which serves the handling of problems concerning the missing, false or inconsistent responses identified during data-processing, editing. Imputation takes place after the data collecting, data entry, editing, consistency checking, and outlier filtering phases. The quality of all these phases can affect all the forthcoming processing phases, including imputation. Quality components affected by imputation  Accuracy: at origin imputation is used to improve the accuracy and completness of the data, but obviously, as a part of the processing it can cause errors. For example: Imputed data usually exhibit less variability than true data.  Timeliness: If we introduce an imputation method then the data capturing phase needs to be shorter. This ha an impact on the quality of the data capture phase. For example, if we plan to use the imputation we do not necessarily need to re- contact data suppliers for the missing or wrong data restoration.

Q2010, Helsinki Imputation indicators  Imputation rate, item level  Imputation rate, unit level  Rate of correction with previously imputed data Calculation: number of cells corrected with previously imputed data / number of imputed cells. Metadata: Data used for imputation with flags, imputed dataset with flags, description of the imputation method. Comment: In case of using several data for the correction of a cell and one of them is corrected, than to imputation with previously imputed data must be considered.  Imputation ratio, item level  Change in standard error due to imputation in case of main indicators

Q2010, Helsinki The next steps of the project  Analysis and assessment of the measuring capability of the above identified indicators  Testing in the case of specific surveys  EU-SILC  Integrated business statistic survey

Q2010, Helsinki Thank you for your attention!