Presentation is loading. Please wait.

Presentation is loading. Please wait.

INFO 7470 Statistical Tools: Edit and Imputation

Similar presentations


Presentation on theme: "INFO 7470 Statistical Tools: Edit and Imputation"— Presentation transcript:

1 INFO 7470 Statistical Tools: Edit and Imputation
John M. Abowd and Lars Vilhuber April 11, 2016 (updated November 3, 2017)

2 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Outline Why learn about edit and imputation procedures Formal models of edits and imputations Missing data overview Missing records Frame or census Survey Missing items Overview of different products Overview of methods Formal multiple imputation methods Examples April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

3 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Why? Users of public-use data can normally identify the existence and consequences of edits and imputations, but don’t have access to data that would improve them Users of restricted-access data normally encounter raw files that require sophisticated edit and imputation procedures in order to use effectively Users of integrated (linked) data from multiple sources face these problems in their most extreme form April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

4 Formal Edit and Imputation Models
Original work on this subject can be found in Fellegi and Holt (1976) One of Fellegi’s many seminal contributions Recent work by Winkler (2008) Formal models distinguish between edits (based on expert judgments) and imputations (based on modeling) State of the art: Manrique-Vallier and Reiter (2014) and Murray and Reiter (JASA, forthcoming) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

5 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Definition of “Edit” Checking each field (variable) of a data record to ensure that it contains a valid entry Examples: NAICS code in range; 0≤age<120 Checking the entries of specified fields (variables) to ensure that they are consistent with each other Examples: job creations – job destructions = accessions – separations; age = data_reference_date – birth_date April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

6 Options When There Is an Edit Failure
Check original files (written questionnaires, census forms, original administrative records) Contact reference entity (household, person, business, establishment) Clerically “correct” data (using expert judgment, not model-base) Automatic edit (using a specified algorithm) Delete record from analysis/tabulation April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

7 Direct Edits Are Expensive
Even when feasible, it is expensive to cross-check every edit failure with original source material It is extremely expensive to re-contact sources, although some re-contacts are built into data collection budgets Computer-assisted survey/census methods can do these edits at the time of the original data collection Re-contacting or revisiting original source material is usually either infeasible or prohibited for administrative records in statistical use April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

8 Example: Age on the SIPP
The SIPP collects both age and birth date (as do the decennial census and ACS) These are often inconsistent; an edit is applied to make them consistent When linked to SSA administrative record data, the birth date on the SSN application becomes available It is often inconsistent with the respondent report And both birth dates are often inconsistent with the observed retirement benefit status of a respondent (also in administrative data) Why? April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

9 The SSA Birth Date Is Respondent Provided until Benefits are Claimed
Prior to 1986, most Americans received SSNs when they entered the labor force not before their first birthday, as is now the case The birth date on SSN application records is “respondent” provided (and not edited by SSA) Further updates to the SSA “Numident” file, which is the master data base of SSNs, are also “respondent provided” Only when a person claims age-dependent benefits (Social Security at 62 or Medicare at 65) does SSA get a birth certificate and apply a true edit to the birth date April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

10 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Lesson Editing, even informal editing, always involves building models even if they are difficult to lay out explicitly Formal edit models involve specifying all the logical relations among the variables (including impossibilities), then imposing them on a probability model like the ones we will consider in this lecture When an edit is made with probability 1 in such a system, the designers of the edit have declared that the expert’s prior judgment cannot be overturned by data In resolving the SIPP/SSA birth date example, the experts declared that the since benefit recipiency was based on audited (via birth certificates) birth dates, it should be taken as “true;” all other values were edited to agree. Note: still don’t have enough information to apply this edit on receipt of SIPP data Doesn’t help for those too young to claim age-eligible benefits. April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

11 Edits and Missing Data Overview
Missing data are a constant feature of both sampling frames (derived from censuses) and surveys Two important types are distinguished Missing record (frame) or interview (survey) Missing item (in either context) Methods differ depending upon type April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

12 Missing Records: Frame or Census
The problem of missing records in a census or sampling frame is detection By definition in these contexts the problem requires external information to solve April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

13 Census of Population and Housing
End-to-end census test (usually two years before) Pre-census housing list review Census processing of housing units found on a block not present on the initial list Post-census evaluation survey Post-census coverage studies April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

14 Economic Censuses and the Business Register
Discussed in lecture 4 Start with tax records Unduplication in the Business Register Weekly updates Multi-units updated with Report of Organization Survey Multi-units discovered during the inter-censal surveys are added to the BR April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

15 Missing Records: Survey
Non-response in a survey is normally handled within the sample design Follow-up (up to a limit) to obtain interview/data Assessment of non-response within sample strata Adjustment of design weights to reflect non-responses April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

16 Missing and/or Inconsistent Items
Edits during the interview (CAPI/CATI/Web form) and during post-processing of the survey Edit or imputation based on the other data in the interview/case (relational edit or imputation) Imputation based on related information on the same respondent, household, or business unit owner Longitudinal: earlier interviews Relational: other members of household, other establishments in the business unit Imputation based on statistical modeling Hot deck Cold deck Model-based: multiple imputation, and other model-based April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

17 Census 2000 PUMS Missing Data
Pre-edit: When the original entry was rejected because it fell outside the range of acceptable values Consistency: Edited or imputed missing characteristics based on other information recorded for the person or housing unit Hot Deck: Supplied the missing information from the record of another person or housing unit. Cold Deck: Supplied missing information from a predetermined distribution See allocation flags for details April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

18 American Community Survey (Current Procedures)
ACS methodology chapter 10 Nonresponse to the mail-out solicitation goes to CATI/CAPI Mail responses given initial automated clerical edit; failures sent to failed-edit follow-up (FEFU), which can result in re-contact (TQA, telephone questionnaire assistance) or full CATI/CAPI interview Whether mailed-in, FEFU, TQA, CATI or CAPI all data enter the data capture file at this point April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

19 American Community Survey (Current Procedures) II
Specialized automated coding procedures for edits to race, Hispanic origin, and ancestry (can involve clerical follow-up) Computer-assisted clerical coding for industry and occupation Automated geocoding for place of birth, migration, place of work (can result in clerical follow-up); place of work particularly troublesome (47% failure rate for automated coding) Quality of underlying data is assessed with an “acceptability index”; too low rejects the person record April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

20 American Community Survey (Current Procedures) III
Missing data not corrected via edit are imputed Two primary imputation methods: Assignment: use of other reported data from respondent or household (e.g., sex, race, citizenship) Hot-deck allocation: discussed later in this lecture April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

21 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
CPS Missing Data Relational edit or imputation: use other information in the record to infer value (based on expert judgment) Longitudinal edits: use values from the previous month if present in sample (based on rules, usually) Hot deck: use values from actual respondents whose data are complete for the, relatively few, conditioning variables April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

22 County Business Patterns
The County and Zip code Business Patterns data are published from the Employer Business Register This is important because variables used in these publications are edited to publication standards The primary imputation method is a longitudinal edit April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

23 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Economic Censuses Like demographic products, there are usually both edited and unedited versions of the publication variables in these files Publication variables (e.g., payroll, employment, sales, geography, ownership) have been edited Most recent files include allocation flags to indicate that a publication variable has been edited or imputed Many historical files include variables that have been edited or imputed but do not include the flags April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

24 QWI Missing Data Procedures (LEHD Infrastructure Files)
Individual data Multiple imputation Employer data Relational edit Bi-directional longitudinal edit Single-value imputation Job data Use multiple imputation of individual data Multiple imputation of place of work Use data for each place of work April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

25 BLS National Longitudinal Surveys
Non-responses to the first wave never enter the data Non-responses to subsequent waves are coded as “interview missing” Respondents are not dropped for missing an interview Special procedures are used to fill critical items from missed interviews when the respondent is interviewed again Item non-response is coded as such April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

26 Federal Reserve Survey of Consumer Finances (SCF)
General information on the Survey of Consumer Finances: Missing data and confidentiality protection are handled with the same multiple imputation procedure April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

27 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
SCF Details Survey collects detailed wealth information from an over-sample of wealthy households Item refusals and item non-response are rampant (see Kennickell article 2011) When there is item refusal, interview instrument attempts to get an interval The reported interval is used in the missing data imputation When the response is deemed sensitive enough for confidentiality protection, the response is treated as an item missing (using the same interval model as above) First major survey released with multiple imputation April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

28 Relational Edit or Imputation
Uses information from the same respondent Example: respondent provided age but not birth date. Use age to impute birth date. Example: some members of household have missing race/ethnicity data. Use other members of same household to impute race/ethnicity April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

29 Longitudinal Edit or Imputation
Look at the respondent’s history in the data to get the value Example: respondent’s employment information missing this month. Impute employment information from previous month Example: establishment industry code missing this quarter. Impute industry code from most recently reported code April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

30 Cross Walks and Other Imputations
In business data, converting an activity code (e.g., SIC) to a different activity code (e.g., NAICS) is a form of missing data imputation This was the original motivation for Rubin’s work (1996 review article) using occupation codes In general, the two activity codes are not done simultaneously for the same entity Often these imputations are treated as 1-1 when they are, in fact, many-to-many April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

31 Probabilistic Methods for Cross Walks
Inputs: original codes new codes information for computing Pr[new code | original code, other data] Processing Randomly assign a new code from the appropriate conditional distribution April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

32 The Theory of Model-based Missing Data Methods
General principles Missing at random Weighting procedures Imputation procedures Hot decks Introduction to model-based procedures April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

33 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
General Principles Most of this lecture is taken from Statistical Analysis with Missing Data, 2nd edition, Roderick J. A. Little and Donald B. Rubin (New York: John Wiley & Sons, 2002) The basic insight is that missing data should be modeled using the same probability and statistical tools that are the basis of all data analysis Missing data are not an anomaly to be swept under the carpet They are an integral part of every analysis April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

34 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Missing Data Patterns Univariate non-response Multivariate non-response Monotone General File matching Latent factors, Bayesian parameters April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

35 Missing Data Mechanisms
The complete data are defined as the matrix Y (n  K). The pattern of missing data is summarized by a matrix of indicator variables M (n  K). The data generating mechanism is summarized by the joint distribution of Y and M. © John M. Abowd and Lars Vilhuber 2016, all rights reserved April 11, 2016

36 Missing Completely at Random
In this case the missing data mechanism does not depend upon the data Y. This case is called MCAR. © John M. Abowd and Lars Vilhuber 2016, all rights reserved April 11, 2016

37 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Missing at Random Partition Y into observed and unobserved parts. Missing at random means that the distribution of M depends only on the observed parts of Y. Called MAR. © John M. Abowd and Lars Vilhuber 2016, all rights reserved April 11, 2016

38 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Not Missing at Random If the condition for MAR fails, then we say that the data are not missing at random, NMAR. Censoring and more elaborate behavioral models often fall into this category. April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

39 The Rubin and Little Taxonomy
Analysis of the complete records only Weighting procedures Imputation-based procedures Model-based procedures April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

40 Analysis of Complete Records Only
Assumes that the data are MCAR Only appropriate for small amounts of missing data Used to be common in economics, less so in sociology Now very rare April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

41 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Weighting Procedures Modify the design weights to correct for missing records Provide an item weight (e.g., earnings and income weights in the CPS) that corrects for missing data on that variable. See Bollinger and Hirsch discussion later in lecture See complete case and weighted complete case discussion in Rubin and Little April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

42 Imputation-based Procedures
Missing values are filled-in and the resulting “completed” data are analyzed Hot deck Mean imputation Regression imputation Some imputation procedures (e.g., Rubin’s multiple imputation) are really model-based procedures April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

43 Imputation Based on Statistical Modeling
Hot deck: use the data from related cases in the same survey to impute missing items (usually as a group) Cold deck: use a fixed probability model to impute the missing items Multiple imputation: use the predictive distribution of the missing item, given all the other items, to impute the missing data April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

44 Current Population Survey
Census Bureau imputation procedures: Relational imputation Longitudinal edit Hot Deck allocation procedure Winkler full edit/imputation system In the technical documentation of methodology for the Current Population Survey, the Census Bureau describes three imputation methods they use with CPS data. 1.        Relational Imputation 2.        Longitudinal edit 3.        Hot deck allocation procedure The first two are more edit procedures than statistical imputation procedures. The examples they give for a relational imputation are when race is missing it is taken on another member of the household and imputed to the person. Longitudinal edit involves going back to a previous interview for the person and taking their answer from there. The hot deck allocation procedure assigns a missing value from a record with similar characteristics and in the same geographic region. Hot decks are always defined by age, race, and sex . Other characteristics used in hot decks vary depending on the nature of the question being referenced. April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

45 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
“Hot Deck” Allocation Labor Force Status Employed Unemployed Not in the Labor Force (Thanks to Warren Brown) Allison describes this procedure as non-parametric, which is a term for statistics that do not require a normal distribution. Also called distribution free statistics. Most common examples of non-parametric statistics are median and semi-interquartile range. Median instead of Mean as a measure of central tendency; and Semi-interquartile range as a description of variability instead of variance and standard deviation. Example from CPS Records are processed and answers to Labor Force Status are checked. This is a categorical variable and valid responses are: Employed, Unemployed, Not in the Labor Force April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

46 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
“Hot Deck” Allocation Black Non-Black Male 16 – 24 25+ ID #0062 Female 16-24 Hypothetically, the "hot deck" holds records based on Age X Race X Sex Age: 16-24, 25+ Race: Black, Non-Black Sex: Male, Female There are 8 cells. A White male, aged 65 is processed and has a valid answer for labor force status. He is employed. His record then bumps the record in that cell and occupies that cell in the hot deck. April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

47 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
“Hot Deck” Allocation Black Non-Black Male 16 – 24 ID #3502 ID #1241 25+ ID #8177 ID #0062 Female 16-24 ID #9923 ID #5923 ID #4396 ID #2271 The other cells are filled in a similar manner. Every time a person with complete answers for questions on: Sex, Age, Race and Labor Force Status is processed they bump whoever occupied the “Hot Deck” cell. When a record with missing Labor Force Status comes along, it gets assigned the Labor Force Status of the person whose record is in the appropriate cell. The advantages of the "hot deck" allocation procedure are that answers are random, valid actual responses, for a similar person based on other characteristics, sensitive to differences between geographic regions, April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

48 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
CPS Example Effects of hot-deck imputation of labor force status April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

49 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Public Use Statistics April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

50 Allocated v. Unallocated
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

51 Bollinger and Hirsch CPS Missing Data
Studies the particular assumptions in the CPS hot deck imputer on wage regressions Census Bureau uses too few variables in its hot deck model Inclusion of additional variables improves the accuracy of the missing data models See Bollinger and Hirsch (2006) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

52 Model-based Procedures
A probability model based on p(Y, M) forms the basis for the analysis This probability model is used as the basis for estimation of parameters or effects of interest Some general-purpose model-based procedures are designed to be combined with likelihood functions that are not specified in advance April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

53 Little and Rubin’s Principles
Imputations should be Conditioned on observed variables Multivariate Draws from a predictive distribution Single imputation methods do not provide a means to correct standard errors for estimation error April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

54 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

55 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

56 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

57 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

58 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

59 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

60 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

61 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

62 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

63 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

64 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

65 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

66 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

67 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

68 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

69 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

70 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

71 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

72 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

73 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

74 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

75 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

76 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

77 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

78 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

79 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

80 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

81 Applications to Complicated Data
Computational formulas for MI data Examples of building Multiply-imputed data files April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

82 Computational Formulas
Assume that you want to estimate something as a function of the data Q(Y) Formulas account for missing data contribution to variance © John M. Abowd and Lars Vilhuber 2016, all rights reserved April 11, 2016

83 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Examples Survey of Consumer Finances Quarterly Workforce Indicators April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

84 Survey of Consumer Finances
Codebook description of missing data procedures Sensitive survey because of large very-wealthy oversample (based on IRS list of most wealthy households in the U.S.) The missing data and confidentiality protection procedures are both based on modeling the complex set of choices given to respondents about how much wealth information to reveal When a respondent does not want to provide an answer, bracketed interval choices are presented Missing data are multiply-imputed based on modeling the conditional distribution of the bracketed intervals, given covariates, and ignorability; see Kennickell (2001) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

85 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
How are the QWIs Built? Raw input files: UI wage records QCEW/ES-202 report Decennial census and ACS files SSA-supplied administrative records Census-derived administrative record household address files LEHD geo-coding system Processed data files: Individual characteristics Employer characteristics Employment history with earnings April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

86 Processing the Input Files
Each quarter the complete history of every individual, every establishment, and every job is processed through the production system Missing data on the individuals are multiply imputed at the national level, posterior predictive distribution is stored Missing data on the employment history record are multiply imputed each quarter from fresh posterior predictive distribution Missing data on the employer characteristics are singly-imputed (explanation to follow) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

87 Examples of Missing Data Problems
Missing demographic data on the national individual file (birth date, sex, race, ethnicity, place of residence, and education) Multiple imputations using information from the individual, establishment, and employment history files Model estimation component updated irregularly Imputations performed once for those in estimation universe, then once when a new PIK is encountered in the production system This process was used on the current QWI and for the S2011 snapshot April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

88 A Very Difficult Missing Data Problem
The employment history records only code employer to the UI account level Establishment characteristics (industry, geo-codes) are missing for multi-unit establishments The establishment (within UI account) is multiply imputed using a dynamic multi-stage probability model Estimation of the posterior predictive distribution depends on the existence of a state with establishments coded on the UI wage record (MN) April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

89 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
How Is It Done? Every quarter the QWI processes over 6 billion employment histories (unique person-employer pair) covering 1990 to 2015 Approximately 30-40% of these histories require multiple employer imputations So, the system does more than 25 billion full information imputations every quarter The information used for the imputations is current, it includes all of the historical information for the person and every establishment associated with that person’s UI account April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

90 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
Does It Work? Full assessment of total jobs, beginning-of-quarter employment, full-quarter employment, monthly earnings of full-quarter employed, total payroll Older assessment using the state that codes both (MN) Summary slide follows April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

91 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

92 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

93 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

94 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

95 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

96 © John M. Abowd and Lars Vilhuber 2016, all rights reserved
April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved

97 Cumulative Effect of All QWI Edits and Imputations
Average Z-scores Missingness Rates Sample Size Entity Size* Average Employment Beginning Employment (b) Full-quarter Employment (f) Earnings (z_w3) all 437 8.79 8.09 10.15 27.1% 31.2% 237,741 1-9 4 1.60 1.46 2.99 33.1% 33.3% 43.6% 95,520 10-99 35 4.84 4.40 6.69 24.4% 25.7% 84,621 160 11.08 10.14 13.52 21.8% 21.6% 20.5% 21,187 354 16.66 15.29 19.37 20.9% 19.1% 11,972 707 23.59 21.68 26.03 20.7% 20.6% 17.9% 8,787 1000+ 5538 56.67 52.61 52.11 20.2% 20.1% 16.1% 15,654 *Entity is county x NAICS sector x race x ethnicity for 2008:q3. Z-score is the ratio of the QWI estimate to the square root of its total variation (within and between implicate components) Missingness rate is the ratio of the between variance to the total variance April 11, 2016 © John M. Abowd and Lars Vilhuber 2016, all rights reserved


Download ppt "INFO 7470 Statistical Tools: Edit and Imputation"

Similar presentations


Ads by Google