Lecture 3 HSPM J716. Efficiency in an estimator Efficiency = low bias and low variance Unbiased with high variance – not very useful Biased with low variance.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Correlation and Regression
Objectives (BPS chapter 24)
Simple Linear Regression
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
What makes one estimator better than another Estimator is jargon term for method of estimating.
Chapter 10 Simple Regression.
Statistics: Data Analysis and Presentation Fr Clinic II.
Additional Topics in Regression Analysis
Lecture 3 HSPM J716. New spreadsheet layout Coefficient Standard error T-statistic – Coefficient ÷ its Standard error.
The Simple Regression Model
T-test.
Chapter Topics Types of Regression Models
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Simple Linear Regression Analysis
Introduction to Probability and Statistics Linear Regression and Correlation.
Ch. 14: The Multiple Regression Model building
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
BCOR 1020 Business Statistics
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Active Learning Lecture Slides
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Inference for regression - Simple linear regression
Chapter 11 Simple Regression
Hypothesis Testing in Linear Regression Analysis
Understanding Multivariate Research Berry & Sanders.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
BPS - 3rd Ed. Chapter 211 Inference for Regression.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
1 MF-852 Financial Econometrics Lecture 10 Serial Correlation and Heteroscedasticity Roy J. Epstein Fall 2003.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Five Ending Wednesday, September 26 (Note: Exam 1 is on September 27)
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Inference for Regression Chapter 14. Linear Regression We can use least squares regression to estimate the linear relationship between two quantitative.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Multiple Regression BPS chapter 28 © 2006 W.H. Freeman and Company.
Scatterplots & Regression Week 3 Lecture MG461 Dr. Meredith Rolfe.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Simple linear regression Tron Anders Moger
Chapter 22: Building Multiple Regression Models Generalization of univariate linear regression models. One unit of data with a value of dependent variable.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Regression Analysis1. 2 INTRODUCTION TO EMPIRICAL MODELS LEAST SQUARES ESTIMATION OF THE PARAMETERS PROPERTIES OF THE LEAST SQUARES ESTIMATORS AND ESTIMATION.
Dependent (response) Variable Independent (control) Variable Random Error XY x1x1 y1y1 x2x2 y2y2 …… xnxn ynyn Raw data: Assumption:  i ‘s are independent.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Multiple Regression Chapter 14.
Lecturer: Ing. Martina Hanová, PhD.. Regression analysis Regression analysis is a tool for analyzing relationships between financial variables:  Identify.
BPS - 5th Ed. Chapter 231 Inference for Regression.
The Simple Linear Regression Model: Specification and Estimation  Theory suggests many relationships between variables  These relationships suggest that.
Inference for Least Squares Lines
Modern Languages Projection Booth Screen Stage Lecturer’s desk broken
BPK 304W Correlation.
CHAPTER 26: Inference for Regression
Basic Practice of Statistics - 3rd Edition Inference for Regression
Simple Linear Regression
Presentation transcript:

Lecture 3 HSPM J716

Efficiency in an estimator Efficiency = low bias and low variance Unbiased with high variance – not very useful Biased with low variance -- worthless

A no-variance, reliable estimator? The 0 estimator

Eyeball vs. Least squares for assignment 1 s/StudentLines/StudentLines.html s/StudentLines/StudentLines.html

Hypothesis testing – parallels among the coin toss, card trick, and assignment 1A experiments A statistic calculated from our data A critical value for that statistic calculated theoretically based on a hypothesis about how the data were generated If our statistic were greater than the critical value, we would reject the hypothesis.

Hypothesis testing – all about calculating the probability of what you got and drawing an inference With the coin toss experiment – A statistic calculated from our data Counted how many tails came up – A critical value for that statistic calculated theoretically based on the hypothesis that the coin was fair 5 consecutive results that are all the same – When our statistic was greater than the critical value, we rejected the hypothesis

Hypothesis testing – all about calculating the probability of what you got and drawing an inference With the card experiment – A statistic calculated from our data Counted how many times I guessed the card – A critical value for that statistic calculated theoretically based on the hypothesis that the any of 52 cards could come up Even one right guess has a probability less than 0.05, so the critical value is 1. – When our statistic was as big as the critical value, we rejected the hypothesis

T statistic hypothesis tests calculate a probability and draw an inference With the assignment 1A spreadsheet – A statistic calculated from our data The estimated coefficient divided by its standard error – A critical value for that statistic calculated theoretically based on the hypothesis that the true line’s slope is – When our statistic is greater than the critical value, we reject the hypothesis

Not rejecting a false hypothesis Type II error in assignment 1A part 2

How the assumptions apply to the eyeball line and the least squares line

Assumption 1 is that there is a true line and that what you see differs from the true line because of random errors up or down for each point. Eyeball line: It's why you drew a line through the points, instead of using a curve or a wiggly line that goes from one point to the next. Least squares: It’s why you built a spreadsheet that calculates the slope and intercept of a line.

Assumption 2 is that the errors have an expected value of 0. Eyeball line: it's why you try to draw the line through the middle of the points, rather than off to one side or tilting differently. Least squares: The average of the residuals is 0. (The residuals are your estimates of the errors.)

Assumption 3 is that the errors all have the same variance. Eyeball line: It's why you don't favor one point over another in drawing the line. Least squares: The spreadsheet’s sum and average rows are simples sums and averages. No data row gets a different weight from another.

Assumption 4 is that the errors are independent, not correlated with each other. Eyeball line: It's why you predict for X=800 using a point on the line Least squares: Its why you predict for X=800 with 800*slope + intercept.

Confidence interval for a coefficient Coefficient ± its standard error × t from table 95% probability that the true coefficient is in the 95% confidence interval? If you do a lot of studies, you can expect that, for 95% of them, the true coefficient will be in the 95% confidence interval. If 0 is in the confidence interval, then the coefficient is not significant.

Assignment 2 All regression results are the same Graphs differ Need reason to use or doubt least squares prediction The reason is in the form of rejecting one or more of the assumptions

Durbin-Watson statistic Serial correlation – Finds significant pattern for clinic 2

Confidence interval for prediction The hyperbolic outline

Formal outlier test? Use confidence interval of prediction – With and without the suspect point? How do you predict when your data have an outlier? – Totally ignoring it seems wrong. – So does letting it sway your results too much. – Investigate and use judgment.

Multiple regression 3 or more dimensions 2 or more X variables Y = α + βX + γZ + error Y = α + β 1 X 1 + β 2 X 2 + … + β p X p + error

Fitting a plane in 3D space Linear assumption – Now a flat plane – The effect of a change in X 1 on Y is the same at all levels of X 1 and X 2 and any other X variables. Residuals are vertical distances from the plane to the data points floating in space.

Multiple regression Separating effects – Example from literature – Example from handout

β interpretation in Y = α + βX + γZ + error β is the effect on Y of changing X by 1, holding Z constant. When X is one unit bigger than you would predict it to be from what Z is, then we expect Y to be β more than what you would predict it would be from what Z is. – Those prediction are based on linear relationships.

β -hat formula

LS Spreadsheet as front end Word processor as back end Interpretation of results – Coefficients – Standard errors – T-statistics – P-values Prediction