Topic 3: Simple Linear Regression. Outline Simple linear regression model –Model parameters –Distribution of error terms Estimation of regression parameters.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Topic 12: Multiple Linear Regression
The Simple Regression Model
Topic 9: Remedies.
Chapter 12 Simple Linear Regression
Regresi Linear Sederhana Pertemuan 01 Matakuliah: I0174 – Analisis Regresi Tahun: Ganjil 2007/2008.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Simple Linear Regression. Start by exploring the data Construct a scatterplot  Does a linear relationship between variables exist?  Is the relationship.
Multiple regression analysis
The Simple Linear Regression Model: Specification and Estimation
Reading – Linear Regression Le (Chapter 8 through 8.1.6) C &S (Chapter 5:F,G,H)
Statistics for the Social Sciences
Chapter 10 Simple Regression.
The Simple Regression Model
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Chapter Topics Types of Regression Models
1 Econ 240A Power 7. 2 Last Week §Normal Distribution §Lab Three: Sampling Distributions §Interval Estimation and HypothesisTesting.
Example: Price Analysis for Diamond Rings in Singapore Variables Response Variable: Price in Singapore dollars (Y) Explanatory Variable: Weight of diamond.
This Week Continue with linear regression Begin multiple regression –Le 8.2 –C & S 9:A-E Handout: Class examples and assignment 3.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Topic 16: Multicollinearity and Polynomial Regression.
© 2002 Prentice-Hall, Inc.Chap 14-1 Introduction to Multiple Regression Model.
Autocorrelation Outline 1) What is it?
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Topic 2: An Example. Leaning Tower of Pisa Construction began in 1173 and by 1178 (2 nd floor), it began to sink Construction resumed in To compensate.
1 Experimental Statistics - week 10 Chapter 11: Linear Regression and Correlation.
Topic 7: Analysis of Variance. Outline Partitioning sums of squares Breakdown degrees of freedom Expected mean squares (EMS) F test ANOVA table General.
Lecture 4 SIMPLE LINEAR REGRESSION.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Lecture 3: Inference in Simple Linear Regression BMTRY 701 Biostatistical Methods II.
Topic 14: Inference in Multiple Regression. Outline Review multiple linear regression Inference of regression coefficients –Application to book example.
Introduction to Linear Regression
Copyright © Cengage Learning. All rights reserved. 12 Simple Linear Regression and Correlation
Topic 17: Interaction Models. Interaction Models With several explanatory variables, we need to consider the possibility that the effect of one variable.
Chapter 14 Inference for Regression AP Statistics 14.1 – Inference about the Model 14.2 – Predictions and Conditions.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
6-1 Introduction To Empirical Models Based on the scatter diagram, it is probably reasonable to assume that the mean of the random variable Y is.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Topic 6: Estimation and Prediction of Y h. Outline Estimation and inference of E(Y h ) Prediction of a new observation Construction of a confidence band.
Topic 13: Multiple Linear Regression Example. Outline Description of example Descriptive summaries Investigation of various models Conclusions.
Topic 25: Inference for Two-Way ANOVA. Outline Two-way ANOVA –Data, models, parameter estimates ANOVA table, EMS Analytical strategies Regression approach.
Topic 26: Analysis of Covariance. Outline One-way analysis of covariance –Data –Model –Inference –Diagnostics and rememdies Multifactor analysis of covariance.
Simple Linear Regression. Data available : (X,Y) Goal : To predict the response Y. (i.e. to obtain the fitted response function f(X)) Least Squares Fitting.
Topic 20: Single Factor Analysis of Variance. Outline Analysis of Variance –One set of treatments (i.e., single factor) Cell means model Factor effects.
© 2001 Prentice-Hall, Inc.Chap 13-1 BA 201 Lecture 18 Introduction to Simple Linear Regression (Data)Data.
Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables.
1 Statistics 262: Intermediate Biostatistics Regression Models for longitudinal data: Mixed Models.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
1 Experimental Statistics - week 12 Chapter 11: Linear Regression and Correlation Chapter 12: Multiple Regression.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Statistics 350 Review. Today Today: Review Simple Linear Regression Simple linear regression model: Y i =  for i=1,2,…,n Distribution of errors.
1 Experimental Statistics - week 11 Chapter 11: Linear Regression and Correlation.
Simple linear regression. What is simple linear regression? A way of evaluating the relationship between two continuous variables. One variable is regarded.
1 Linear Regression Model. 2 Types of Regression Models.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
The simple linear regression model and parameter estimation
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
Linear Regression.
Simple Linear Regression - Introduction
6-1 Introduction To Empirical Models
Statistical Assumptions for SLR
Linear regression Fitting a straight line to observations.
The Simple Linear Regression Model: Specification and Estimation
Chapter 14 Inference for Regression
Topic 11: Matrix Approach to Linear Regression
Regression Models - Introduction
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Topic 3: Simple Linear Regression

Outline Simple linear regression model –Model parameters –Distribution of error terms Estimation of regression parameters –Method of least squares –Maximum likelihood

Data for Simple Linear Regression Observe i=1,2,...,n pairs of variables Each pair often called a case Y i = i th response variable X i = i th explanatory variable

Simple Linear Regression Model Y i = b 0 + b 1 X i + e i b 0 is the intercept b 1 is the slope e i is a random error term –E( e i )=0 and s 2 ( e i ) =s 2 –e i and e j are uncorrelated

Simple Linear Normal Error Regression Model Y i = b 0 + b 1 X i + e i b 0 is the intercept b 1 is the slope e i is a Normally distributed random error with mean 0 and variance σ 2 e i and e j are uncorrelated → indep

Model Parameters β 0 : the intercept β 1 : the slope σ 2 : the variance of the error term

Features of Both Regression Models Y i = β 0 + β 1 X i + e i E (Y i ) = β 0 + β 1 X i + E(e i ) = β 0 + β 1 X i Var(Y i ) = 0 + var(e i ) = σ 2 –Mean of Y i determined by value of X i –All possible means fall on a line –The Y i vary about this line

Features of Normal Error Regression Model Y i = β 0 + β 1 X i + e i If e i is Normally distributed then Y i is N(β 0 + β 1 X i, σ 2 ) (A.36) Does not imply the collection of Y i are Normally distributed

Fitted Regression Equation and Residuals Ŷ i = b 0 + b 1 X i –b 0 is the estimated intercept –b 1 is the estimated slope e i : residual for i th case e i = Y i – Ŷ i = Y i – (b 0 + b 1 X i )

X=82 Ŷ 82 =b 0 + b 1 82 e 82 =Y 82 -Ŷ 82

Plot the residuals proc gplot data=a2; plot resid*year vref=0; where lean ne.; run; Continuation of pisa.sas Using data set from output statement vref=0 adds horizontal line to plot at zero

e 82

Least Squares Want to find “best” b 0 and b 1 Will minimize Σ(Y i – (b 0 + b 1 X i ) ) 2 Use calculus: take derivative with respect to b 0 and with respect to b 1 and set the two resulting equations equal to zero and solve for b 0 and b 1 See KNNL pgs 16-17

Least Squares Solution These are also maximum likelihood estimators for Normal error model, see KNNL pp 30-32

Maximum Likelihood

Estimation of σ 2

Standard output from Proc REG Analysis of Variance SourceDF Sum of Squares Mean SquareF ValuePr > F Model <.0001 Error Corrected Total Root MSE R-Square Dependent Mean Adj R-Sq Coeff Var MSE df e s

Properties of Least Squares Line The line always goes through Other properties on pgs 23-24

Background Reading Chapter 1 –1.6 : Estimation of regression function –1.7 : Estimation of error variance –1.8 : Normal regression model Chapter 2 –2.1 and 2.2 : inference concerning  ’s Appendix A –A.4, A.5, A.6, and A.7