Chicago Insurance Redlining Example Were insurance companies in Chicago denying insurance in neighborhoods based on race?

Slides:



Advertisements
Similar presentations
Multicollinearity.
Advertisements

More on understanding variance inflation factors (VIFk)
1 Outliers and Influential Observations KNN Ch. 10 (pp )
BA 275 Quantitative Business Methods
Objectives (BPS chapter 24)
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Simple Linear Regression Estimates for single and mean responses.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Analysis of Economic Data
Statistical Analysis of the Regression Point Displacement Design (RPD)
Applied Linear Regression CSTAT Workshop March 16, 2007 Vince Melfi.
Some more issues of time series analysis Time series regression with modelling of error terms In a time series regression model the error terms are tentatively.
Note 14 of 5E Statistics with Economics and Business Applications Chapter 12 Multiple Regression Analysis A brief exposition.
Pris mot antal rum. Regression Analysis: Price versus Rooms The regression equation is Price = Rooms Predictor Coef SE Coef T P Constant.
REGRESSION AND CORRELATION
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Descriptive measures of the strength of a linear association r-squared and the (Pearson) correlation coefficient r.
Hypothesis tests for slopes in multiple linear regression model Using the general linear test and sequential sums of squares.
Variable selection and model building Part II. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Chapter 12: Linear Regression 1. Introduction Regression analysis and Analysis of variance are the two most widely used statistical procedures. Regression.
Name: Angelica F. White WEMBA10. Teach students how to make sound decisions and recommendations that are based on reliable quantitative information During.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 12-1 Correlation and Regression.
Introduction to Linear Regression
Understanding Variation in Patient Satisfaction. All Rights Reserved, Juran Institute, Inc. Understanding Variation in Patient Satisfaction 2.PPT Variation.
TEAS prep course trends YOUR FUTURE BEGINS TODAY Joel Collazo, MD Maria E Guzman, MPM.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
An alternative approach to testing for a linear association The Analysis of Variance (ANOVA) Table.
Detecting and reducing multicollinearity. Detecting multicollinearity.
Warm-up Ch.11 Inference for Linear Regression Day 2 1. Which of the following are true statements? I. By the Law of Large Numbers, the mean of a random.
Copyright ©2011 Nelson Education Limited Linear Regression and Correlation CHAPTER 12.
Solutions to Tutorial 5 Problems Source Sum of Squares df Mean Square F-test Regression Residual Total ANOVA Table Variable.
Obesity in America Nina Netupsky. Prevalence of Obesity in America  Over the past 30 years the number of individuals that have been diagnosed with obesity.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
14- 1 Chapter Fourteen McGraw-Hill/Irwin © 2006 The McGraw-Hill Companies, Inc., All Rights Reserved.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Non-linear Regression Example.
Lack of Fit (LOF) Test A formal F test for checking whether a specific type of regression function adequately fits the data.
Multiple regression. Example: Brain and body size predictive of intelligence? Sample of n = 38 college students Response (Y): intelligence based on the.
Fitting Curves to Data 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 5: Fitting Curves to Data Terry Dielman Applied Regression.
Overview of our study of the multiple linear regression model Regression models with more than one slope parameter.
Missing Girls: An Evaluation of Sex Ratios in China As humanitarian and gender equality issues have been gaining more attention in the last several decades,
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Copyright © 2004 by The McGraw-Hill Companies, Inc. All rights reserved.
Business Statistics, 4e, by Ken Black. © 2003 John Wiley & Sons Business Statistics, 4e by Ken Black Chapter 14 Multiple Regression Analysis.
Statistics and Numerical Method Part I: Statistics Week VI: Empirical Model 1/2555 สมศักดิ์ ศิวดำรงพงศ์ 1.
Inference for regression - More details about simple linear regression IPS chapter 10.2 © 2006 W.H. Freeman and Company.
732G21/732G28/732A35 Lecture 4. Variance-covariance matrix for the regression coefficients 2.
Multicollinearity. Multicollinearity (or intercorrelation) exists when at least some of the predictor variables are correlated among themselves. In observational.
Interaction regression models. What is an additive model? A regression model with p-1 predictor variables contains additive effects if the response function.
Agenda 1.Exam 2 Review 2.Regression a.Prediction b.Polynomial Regression.
732G21/732G28/732A35 Lecture 6. Example second-order model with one predictor 2 Electricity consumption (Y)Home size (X)
Chapter 12 Inference on the Least-squares Regression Line; ANOVA 12.3 One-way Analysis of Variance.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
Descriptive measures of the degree of linear association R-squared and correlation.
Design and Analysis of Experiments (5) Fitting Regression Models Kyung-Ho Park.
Analysis of variance approach to regression analysis … an (alternative) approach to testing for a linear association.
1 Multiple Regression. 2 Model There are many explanatory variables or independent variables x 1, x 2,…,x p that are linear related to the response variable.
Model selection and model building. Model selection Selection of predictor variables.
David Housman for Math 323 Probability and Statistics Class 05 Ion Sensitive Electrodes.
Announcements There’s an in class exam one week from today (4/30). It will not include ANOVA or regression. On Thursday, I will list covered material and.
Chapter 20 Linear and Multiple Regression
Introduction to Regression Lecture 6.2
Chapter 13 Created by Bethany Stubbe and Stephan Kogitz.
Inference for Regression Lines
Solutions for Tutorial 3
Solutions of Tutorial 10 SSE df RMS Cp Radjsq SSE1 F Xs c).
Residuals The residuals are estimate of the error
Solution 9 1. a) From the matrix plot, 1) The assumption about linearity seems ok; 2).The assumption about measurement errors can not be checked at this.
Multiple Regression Chapter 14.
Solutions of Tutorial 9 SSE df RMS Cp Radjsq SSE1 F Xs c).
Chapter Fourteen McGraw-Hill/Irwin
Presentation transcript:

Chicago Insurance Redlining Example Were insurance companies in Chicago denying insurance in neighborhoods based on race?

The background In some US cities, services such as insurance are denied based on race This is sometimes called “redlining.” For insurance, many states have a “FAIR” plan available, for (and limited to) those who cannot obtain insurance in the regular market. So an area with high numbers of FAIR plan policies is an area where it is hard to get insurance in the regular market.

The data (for 47 zip codes near Chicago) involact = # of new FAIR plan policies and renewals per 100 housing units race = % minority theft = theft per 1000 population fire = fires per 100 housing units income = median family income in $1000s

First, some description Descriptive statistics for the variables Box plots Histograms Matrix plots etc.

Descriptive Statistics: race, fire, theft, age, involact, income Variable N N* Mean SE Mean StDev Minimum Q1 Median Q3 race fire theft age involact income Variable Maximum race fire theft age involact income

Simple linear regression model Fit a model with involact as the response and race as the predictor A strong positive relationship gives some evidence for redlining

What’s next The matrix plot showed that race is correlated with other predictors, e.g., income, fire, etc. So it’s possible that these are the important factors in influencing involact Next the full model is fit

The regression equation is involact = race fire theft age income Predictor Coef SE Coef T P Constant race fire theft age income

S = R-Sq = 75.1% R-Sq(adj) = 72.0% Analysis of Variance Source DF SS MS F P Regression Residual Error Total

What have we learned? Race is still highly significant (t = 3.94, p-value ≈ 0) in the full model Income is not significant (this isn’t surprising, since race and income are highly correlated).

Diagnostics Some plots are next. Uninteresting (good!) We’ll ignore more substantial diagnostics such as looking at leverage and influence, although these should be done.

Model selection Response is involact i t n r f h c a i e a o Mallows c r f g m Vars R-Sq R-Sq(adj) Cp S e e t e e X X X X X X X X X X X X X X X