Download presentation
Presentation is loading. Please wait.
Published byAntony Montgomery Modified over 8 years ago
1
INFO 7470 Statistical Tools: Edit and Imputation John M. Abowd and Lars Vilhuber April 11, 2016
2
Outline Why learn about edit and imputation procedures Formal models of edits and imputations Missing data overview Missing records – Frame or census – Survey Missing items Overview of different products Overview of methods Formal multiple imputation methods Examples April 11, 20162 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
3
Why? Users of public-use data can normally identify the existence and consequences of edits and imputations, but don’t have access to data that would improve them Users of restricted-access data normally encounter raw files that require sophisticated edit and imputation procedures in order to use effectively Users of integrated (linked) data from multiple sources face these problems in their most extreme form April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 3
4
Formal Edit and Imputation Models Original work on this subject can be found in Fellegi and Holt (1976)1976 One of Fellegi’s many seminal contributionsFellegi’s Recent work by Winkler (2008)2008 Formal models distinguish between edits (based on expert judgments) and imputations (based on modeling) State of the art: Manrique-Vallier and Reiter (2014) and Murray and Reiter (JASA, forthcoming)2014forthcoming April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 4
5
Definition of “Edit” Checking each field (variable) of a data record to ensure that it contains a valid entry – Examples: NAICS code in range; 0≤age<120 Checking the entries of specified fields (variables) to ensure that they are consistent with each other – Examples: – job creations – job destructions = accessions – separations; – age = data_reference_date – birth_date April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 5
6
Options When There Is an Edit Failure 1.Check original files (written questionnaires, census forms, original administrative records) 2.Contact reference entity (household, person, business, establishment) 3.Clerically “correct” data (using expert judgment, not model-base) 4.Automatic edit (using a specified algorithm) 5.Delete record from analysis/tabulation April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 6
7
Direct Edits Are Expensive Even when feasible, it is expensive to cross-check every edit failure with original source material It is extremely expensive to re-contact sources, although some re-contacts are built into data collection budgets Computer-assisted survey/census methods can do these edits at the time of the original data collection Re-contacting or revisiting original source material is usually either infeasible or prohibited for administrative records in statistical use April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 7
8
Example: Age on the SIPP The SIPP collects both age and birth date (as do the decennial census and ACS) These are often inconsistent; an edit is applied to make them consistent When linked to SSA administrative record data, the birth date on the SSN application becomes available It is often inconsistent with the respondent report And both birth dates are often inconsistent with the observed retirement benefit status of a respondent (also in administrative data) Why? April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 8
9
The SSA Birth Date Is Respondent Provided until Benefits are Claimed Prior to 1986, most Americans received SSNs when they entered the labor force not before their first birthday, as is now the case The birth date on SSN application records is “respondent” provided (and not edited by SSA) Further updates to the SSA “Numident” file, which is the master data base of SSNs, are also “respondent provided” Only when a person claims age-dependent benefits (Social Security at 62 or Medicare at 65) does SSA get a birth certificate and apply a true edit to the birth date April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 9
10
Lesson Editing, even informal editing, always involves building models even if they are difficult to lay out explicitly Formal edit models involve specifying all the logical relations among the variables (including impossibilities), then imposing them on a probability model like the ones we will consider in this lecture When an edit is made with probability 1 in such a system, the designers of the edit have declared that the expert’s prior judgment cannot be overturned by data In resolving the SIPP/SSA birth date example, the experts declared that the since benefit recipiency was based on audited (via birth certificates) birth dates, it should be taken as “true;” all other values were edited to agree. – Note: still don’t have enough information to apply this edit on receipt of SIPP data – Doesn’t help for those too young to claim age-eligible benefits. April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 10
11
Edits and Missing Data Overview Missing data are a constant feature of both sampling frames (derived from censuses) and surveys Two important types are distinguished – Missing record (frame) or interview (survey) – Missing item (in either context) Methods differ depending upon type April 11, 201611 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
12
Missing Records: Frame or Census The problem of missing records in a census or sampling frame is detection By definition in these contexts the problem requires external information to solve April 11, 201612 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
13
Census of Population and Housing End-to-end census test (usually two years before) Pre-census housing list review Census processing of housing units found on a block not present on the initial list Post-census evaluation survey Post-census coverage studies April 11, 201613 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
14
Economic Censuses and the Business Register Discussed in lecture 4 Start with tax records Unduplication in the Business Register Weekly updates Multi-units updated with Report of Organization Survey Multi-units discovered during the inter-censal surveys are added to the BR April 11, 201614 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
15
Missing Records: Survey Non-response in a survey is normally handled within the sample design Follow-up (up to a limit) to obtain interview/data Assessment of non-response within sample strata Adjustment of design weights to reflect non- responses April 11, 201615 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
16
Missing and/or Inconsistent Items Edits during the interview (CAPI/CATI/Web form) and during post-processing of the survey Edit or imputation based on the other data in the interview/case (relational edit or imputation) Imputation based on related information on the same respondent, household, or business unit owner – Longitudinal: earlier interviews – Relational: other members of household, other establishments in the business unit Imputation based on statistical modeling – Hot deck – Cold deck – Model-based: multiple imputation, and other model-based April 11, 201616 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
17
Census 2000 PUMS Missing Data Pre-edit: When the original entry was rejected because it fell outside the range of acceptable values Consistency: Edited or imputed missing characteristics based on other information recorded for the person or housing unit Hot Deck: Supplied the missing information from the record of another person or housing unit. Cold Deck: Supplied missing information from a predetermined distribution See allocation flags for details April 11, 201617 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
18
American Community Survey (Current Procedures) ACS methodology chapter 10 Nonresponse to the mail-out solicitation goes to CATI/CAPI Mail responses given initial automated clerical edit; failures sent to failed-edit follow-up (FEFU), which can result in re-contact (TQA, telephone questionnaire assistance) or full CATI/CAPI interview Whether mailed-in, FEFU, TQA, CATI or CAPI all data enter the data capture file at this point April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 18
19
American Community Survey (Current Procedures) II Specialized automated coding procedures for edits to race, Hispanic origin, and ancestry (can involve clerical follow-up) Computer-assisted clerical coding for industry and occupation Automated geocoding for place of birth, migration, place of work (can result in clerical follow-up); place of work particularly troublesome (47% failure rate for automated coding) Quality of underlying data is assessed with an “acceptability index”; too low rejects the person record April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 19
20
American Community Survey (Current Procedures) III Missing data not corrected via edit are imputed Two primary imputation methods: – Assignment: use of other reported data from respondent or household (e.g., sex, race, citizenship) – Hot-deck allocation: discussed later in this lecture April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 20
21
CPS Missing Data Relational edit or imputation: use other information in the record to infer value (based on expert judgment) Longitudinal edits: use values from the previous month if present in sample (based on rules, usually) Hot deck: use values from actual respondents whose data are complete for the, relatively few, conditioning variables April 11, 201621 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
22
County Business Patterns The County and Zip code Business Patterns data are published from the Employer Business Register This is important because variables used in these publications are edited to publication standards The primary imputation method is a longitudinal edit http://www.census.gov/econ/cbp/methodology.htm April 11, 201622 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
23
Economic Censuses Like demographic products, there are usually both edited and unedited versions of the publication variables in these files Publication variables (e.g., payroll, employment, sales, geography, ownership) have been edited Most recent files include allocation flags to indicate that a publication variable has been edited or imputed Many historical files include variables that have been edited or imputed but do not include the flags April 11, 201623 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
24
QWI Missing Data Procedures (LEHD Infrastructure Files) Individual data – Multiple imputation Employer data – Relational edit – Bi-directional longitudinal edit – Single-value imputation Job data – Use multiple imputation of individual data – Multiple imputation of place of work Use data for each place of work April 11, 201624 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
25
BLS National Longitudinal Surveys Non-responses to the first wave never enter the data Non-responses to subsequent waves are coded as “interview missing” Respondents are not dropped for missing an interview – Special procedures are used to fill critical items from missed interviews when the respondent is interviewed again Item non-response is coded as such April 11, 201625 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
26
Federal Reserve Survey of Consumer Finances (SCF) General information on the Survey of Consumer Finances: http://www.federalreserve.gov/pubs/oss/oss2/scfindex.html http://www.federalreserve.gov/pubs/oss/oss2/scfindex.html Missing data and confidentiality protection are handled with the same multiple imputation procedure April 11, 201626 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
27
SCF Details Survey collects detailed wealth information from an over-sample of wealthy households Item refusals and item non-response are rampant (see Kennickell article 2011)2011 When there is item refusal, interview instrument attempts to get an interval The reported interval is used in the missing data imputation When the response is deemed sensitive enough for confidentiality protection, the response is treated as an item missing (using the same interval model as above) First major survey released with multiple imputation April 11, 201627 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
28
Relational Edit or Imputation Uses information from the same respondent Example: respondent provided age but not birth date. Use age to impute birth date. Example: some members of household have missing race/ethnicity data. Use other members of same household to impute race/ethnicity April 11, 201628 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
29
Longitudinal Edit or Imputation Look at the respondent’s history in the data to get the value Example: respondent’s employment information missing this month. Impute employment information from previous month Example: establishment industry code missing this quarter. Impute industry code from most recently reported code April 11, 201629 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
30
Cross Walks and Other Imputations In business data, converting an activity code (e.g., SIC) to a different activity code (e.g., NAICS) is a form of missing data imputation – This was the original motivation for Rubin’s work (1996 review article) using occupation codes1996 review article In general, the two activity codes are not done simultaneously for the same entity Often these imputations are treated as 1-1 when they are, in fact, many-to-many April 11, 201630 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
31
Probabilistic Methods for Cross Walks Inputs: – original codes – new codes – information for computing Pr[new code | original code, other data] Processing – Randomly assign a new code from the appropriate conditional distribution April 11, 201631 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
32
The Theory of Model-based Missing Data Methods General principles Missing at random Weighting procedures Imputation procedures Hot decks Introduction to model-based procedures April 11, 201632 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
33
General Principles Most of this lecture is taken from Statistical Analysis with Missing Data, 2nd edition, Roderick J. A. Little and Donald B. Rubin (New York: John Wiley & Sons, 2002) The basic insight is that missing data should be modeled using the same probability and statistical tools that are the basis of all data analysis Missing data are not an anomaly to be swept under the carpet They are an integral part of every analysis April 11, 201633 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
34
Missing Data Patterns Univariate non-response Multivariate non-response Monotone General File matching Latent factors, Bayesian parameters April 11, 201634 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
35
Missing Data Mechanisms The complete data are defined as the matrix Y (n K). The pattern of missing data is summarized by a matrix of indicator variables M (n K). The data generating mechanism is summarized by the joint distribution of Y and M. April 11, 201635© John M. Abowd and Lars Vilhuber 2016, all rights reserved
36
Missing Completely at Random In this case the missing data mechanism does not depend upon the data Y. This case is called MCAR. April 11, 201636© John M. Abowd and Lars Vilhuber 2016, all rights reserved
37
Missing at Random Partition Y into observed and unobserved parts. Missing at random means that the distribution of M depends only on the observed parts of Y. Called MAR. April 11, 201637© John M. Abowd and Lars Vilhuber 2016, all rights reserved
38
Not Missing at Random If the condition for MAR fails, then we say that the data are not missing at random, NMAR. Censoring and more elaborate behavioral models often fall into this category. April 11, 201638 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
39
The Rubin and Little Taxonomy Analysis of the complete records only Weighting procedures Imputation-based procedures Model-based procedures April 11, 201639 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
40
Analysis of Complete Records Only Assumes that the data are MCAR Only appropriate for small amounts of missing data Used to be common in economics, less so in sociology Now very rare April 11, 201640 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
41
Weighting Procedures Modify the design weights to correct for missing records Provide an item weight (e.g., earnings and income weights in the CPS) that corrects for missing data on that variable. See Bollinger and Hirsch discussion later in lecture See complete case and weighted complete case discussion in Rubin and Little April 11, 201641 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
42
Imputation-based Procedures Missing values are filled-in and the resulting “completed” data are analyzed – Hot deck – Mean imputation – Regression imputation Some imputation procedures (e.g., Rubin’s multiple imputation) are really model-based procedures April 11, 201642 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
43
Imputation Based on Statistical Modeling Hot deck: use the data from related cases in the same survey to impute missing items (usually as a group) Cold deck: use a fixed probability model to impute the missing items Multiple imputation: use the predictive distribution of the missing item, given all the other items, to impute the missing data April 11, 201643 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
44
Current Population Survey Census Bureau imputation procedures: – Relational imputation – Longitudinal edit – Hot Deck allocation procedure – Winkler full edit/imputation system April 11, 201644 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
45
“Hot Deck” Allocation Labor Force Status – Employed – Unemployed – Not in the Labor Force (Thanks to Warren Brown) April 11, 201645 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
46
“Hot Deck” Allocation BlackNon-Black Male 16 – 24 25+ID #0062 Female 16-24 25+ April 11, 201646 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
47
“Hot Deck” Allocation BlackNon-Black Male 16 – 24ID #3502ID #1241 25+ID #8177ID #0062 Female 16-24ID #9923ID #5923 25+ID #4396ID #2271 April 11, 201647 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
48
CPS Example Effects of hot-deck imputation of labor force status April 11, 201648 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
49
Public Use Statistics April 11, 201649 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
50
Allocated v. Unallocated April 11, 201650 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
51
Bollinger and Hirsch CPS Missing Data Studies the particular assumptions in the CPS hot deck imputer on wage regressions Census Bureau uses too few variables in its hot deck model Inclusion of additional variables improves the accuracy of the missing data models See Bollinger and Hirsch (2006)2006 April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 51
52
Model-based Procedures A probability model based on p(Y, M) forms the basis for the analysis This probability model is used as the basis for estimation of parameters or effects of interest Some general-purpose model-based procedures are designed to be combined with likelihood functions that are not specified in advance April 11, 201652 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
53
Little and Rubin’s Principles Imputations should be – Conditioned on observed variables – Multivariate – Draws from a predictive distribution Single imputation methods do not provide a means to correct standard errors for estimation error April 11, 201653 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
54
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 54
55
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 55
56
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 56
57
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 57
58
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 58
59
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 59
60
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 60
61
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 61
62
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 62
63
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 63
64
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 64
65
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 65
66
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 66
67
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 67
68
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 68
69
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 69
70
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 70
71
Applications to Complicated Data Computational formulas for MI data Examples of building Multiply-imputed data files April 11, 201671 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
72
Computational Formulas Assume that you want to estimate something as a function of the data Q(Y) Formulas account for missing data contribution to variance April 11, 201672© John M. Abowd and Lars Vilhuber 2016, all rights reserved
73
Examples Survey of Consumer Finances Quarterly Workforce Indicators April 11, 201673 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
74
Survey of Consumer Finances Codebook description of missing data procedures Sensitive survey because of large very-wealthy oversample (based on IRS list of most wealthy households in the U.S.) The missing data and confidentiality protection procedures are both based on modeling the complex set of choices given to respondents about how much wealth information to reveal When a respondent does not want to provide an answer, bracketed interval choices are presented Missing data are multiply-imputed based on modeling the conditional distribution of the bracketed intervals, given covariates, and ignorability; see Kennickell (2001)2001 April 11, 201674 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
75
How are the QWIs Built? Raw input files: – UI wage records – QCEW/ES-202 report – Decennial census and ACS files – SSA-supplied administrative records – Census-derived administrative record household address files – LEHD geo-coding system Processed data files: – Individual characteristics – Employer characteristics – Employment history with earnings April 11, 201675 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
76
Processing the Input Files Each quarter the complete history of every individual, every establishment, and every job is processed through the production system Missing data on the individuals are multiply imputed at the national level, posterior predictive distribution is stored Missing data on the employment history record are multiply imputed each quarter from fresh posterior predictive distribution Missing data on the employer characteristics are singly- imputed (explanation to follow) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 76
77
Examples of Missing Data Problems Missing demographic data on the national individual file (birth date, sex, race, ethnicity, place of residence, and education) – Multiple imputations using information from the individual, establishment, and employment history files – Model estimation component updated irregularly – Imputations performed once for those in estimation universe, then once when a new PIK is encountered in the production system This process was used on the current QWI and for the S2011 snapshot April 11, 201677 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
78
A Very Difficult Missing Data Problem The employment history records only code employer to the UI account level Establishment characteristics (industry, geo- codes) are missing for multi-unit establishments The establishment (within UI account) is multiply imputed using a dynamic multi-stage probability model Estimation of the posterior predictive distribution depends on the existence of a state with establishments coded on the UI wage record (MN) April 11, 201678 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
79
How Is It Done? Every quarter the QWI processes over 6 billion employment histories (unique person-employer pair) covering 1990 to 2015 Approximately 30-40% of these histories require multiple employer imputations So, the system does more than 25 billion full information imputations every quarter The information used for the imputations is current, it includes all of the historical information for the person and every establishment associated with that person’s UI account April 11, 201679 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
80
Does It Work? Full assessment of total jobs, beginning-of- quarter employment, full-quarter employment, monthly earnings of full-quarter employed, total payroll Older assessment using the state that codes both (MN) Summary slide follows April 11, 201680 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
81
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 81
82
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 82
83
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 83
84
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 84
85
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 85
86
April 11, 201686 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
87
Cumulative Effect of All QWI Edits and Imputations Average Z-scoresMissingness Rates Sample Size Entity Size* Average Employment Beginning Employment (b) Full-quarter Employment (f) Earnings (z_w3) Beginning Employment (b) Full-quarter Employment (f) Earnings (z_w3) all4378.798.0910.1527.1% 31.2%237,741 1-941.601.462.9933.1%33.3%43.6%95,520 10-99354.844.406.6924.4% 25.7%84,621 100-24916011.0810.1413.5221.8%21.6%20.5%21,187 250-49935416.6615.2919.3720.9% 19.1%11,972 500-99970723.5921.6826.0320.7%20.6%17.9%8,787 1000+553856.6752.6152.1120.2%20.1%16.1%15,654 *Entity is county x NAICS sector x race x ethnicity for 2008:q3. Z-score is the ratio of the QWI estimate to the square root of its total variation (within and between implicate components) Missingness rate is the ratio of the between variance to the total variance April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved 87
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.