…Don’t be afraid of others, because they are bigger than you

Slides:



Advertisements
Similar presentations
Regression and correlation methods
Advertisements

Section 10-3 Regression.
The Simple Regression Model
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Basic Statistical Concepts Psych 231: Research Methods in Psychology.
The Simple Regression Model
Basic Statistical Concepts
Statistics Psych 231: Research Methods in Psychology.
SIMPLE LINEAR REGRESSION
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Simple Linear Regression Analysis
Quantitative Business Analysis for Decision Making Simple Linear Regression.
1 Simple Linear Regression Linear regression model Prediction Limitation Correlation.
BCOR 1020 Business Statistics
FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Correlation & Regression
Active Learning Lecture Slides
Regression and Correlation Methods Judy Zhong Ph.D.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
Ch4 Describing Relationships Between Variables. Pressure.
Ch4 Describing Relationships Between Variables. Section 4.1: Fitting a Line by Least Squares Often we want to fit a straight line to data. For example.
1 Everyday is a new beginning in life. Every moment is a time for self vigilance.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
1 Lecture 4 Main Tasks Today 1. Review of Lecture 3 2. Accuracy of the LS estimators 3. Significance Tests of the Parameters 4. Confidence Interval 5.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
LECTURE 9 Tuesday, 24 FEBRUARY STA291 Fall Administrative 4.2 Measures of Variation (Empirical Rule) 4.4 Measures of Linear Relationship Suggested.
Regression Analysis Deterministic model No chance of an error in calculating y for a given x Probabilistic model chance of an error First order linear.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Part II Exploring Relationships Between Variables.
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Chapter 13 Simple Linear Regression
The simple linear regression model and parameter estimation
Regression and Correlation of Data Summary
Regression Analysis AGEC 784.
LECTURE 13 Thursday, 8th October
Correlation and Simple Linear Regression
Inference for Regression
Linear Regression and Correlation Analysis
Chapter 5 STATISTICS (PART 4).
...Relax... 9/21/2018 ST3131, Lecture 3 ST5213 Semester II, 2000/2001
Econ 3790: Business and Economics Statistics
CHAPTER 10 Correlation and Regression (Objectives)
Correlation and Simple Linear Regression
Correlation and Regression
Lecture Notes The Relation between Two Variables Q Q
Simple Linear Regression
Correlation and Simple Linear Regression
Review of Chapter 2 Some Basic Concepts: Sample center
Interpretation of Regression Coefficients
Correlation and Regression
SIMPLE LINEAR REGRESSION
Simple Linear Regression and Correlation
Simple Linear Regression
SIMPLE LINEAR REGRESSION
Ch 4.1 & 4.2 Two dimensions concept
Created by Erin Hodgess, Houston, Texas
Simple Linear Regression
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

…Don’t be afraid of others, because they are bigger than you …Don’t be afraid of others, because they are bigger than you. The real size could be measured in the wisdom. 11/10/2018 ST3131, Lecture 2

Chapter 2 Simple Linear Regression ST5213 Semester II, 2000/2001 Chapter 2 Simple Linear Regression In this Chapter, we consider the simplest regression model : Simple Linear Regression (SLR) Model Which describes the linear relationship between Y and X. Tasks: 1. Review of some basic statistics 2. Define measures of the direction / strength of the linear relationship between Y and X. 3. Find the formulas for the estimators and . 11/10/2018 ST3131, Lecture 2

Review of Some Basic Statistics Let Y and X have n observations as follows: Summary Statistics: Mean : , average of observations of Y, measure for sample center of Y. Deviation: , differences of observation from . Variance: , average of squared deviations of Y. Standard Deviation: , measure for spread of Y. Standardization of Y: Similarly, we can define all terms for X. 11/10/2018 ST3131, Lecture 2

Properties of Standardized Variables Proof: 11/10/2018 ST3131, Lecture 2

Exact Linear Relationship between Y and X Given Y= + X, consider the Linear Relationship between Y and X, Case 1). Positive, X increases, Y increases, when ( ) > 0; 2). Negative, X increases, Y decreases, when ( ) <0; 3). Linearly Uncorrelated, X changes, Y does NOT change, when ( ) =0. =1, = -1 =1, = 0 =1, = 1 Conclusion: ( ) is an Indicator of the direction of Y changing with X. 11/10/2018 ST3131, Lecture 2

Distorted Linear Relationship between Y and X Given Y= + X + , consider the Linear Relationship between Y and X Case 1). Positive, X increases, Y almost increases when ( ) >0; 2). Negative, X increases, Y almost decreases when ( ) <0; 3).Linearly Uncorrelated, X changes, Y almost does NOT change, when ( )=0 =1, = -1 =1, = 0 =1, = -1 Conclusion: ( ) is also an Indicator of the Direction of Y changing with X. 11/10/2018 ST3131, Lecture 2

Intuitive Derivation of the LS-estimators Sample Covariance of Y and X: the summation of cross-products of deviations of Y and X divided by (n-1). Intuitive Derivation: We have , thus we have Thus, 11/10/2018 ST3131, Lecture 2

Formulas for the LS-estimators Assume ( ) =0, and =0. The LS-estimators are Since =0, Cov(Y,X) has the same signs as those of Thus Cov(Y,X) is also an Indicator of the direction of the Linear Relationship between Y and X: Case 1) Positive when Cov(Y,X)>0; 2) Negative when Cov(Y,X)<0; 3) Uncorrelated when Cov(Y,X)=0. Summary: Indicators of the Direction of the Linear Relationship between Y and X 1). The slope 2) The slope estimator 3). Cov (Y,X) 11/10/2018 ST3131, Lecture 2

Properties of Cov(Y,X) 1. Symmetric, i.e., Cov(Y,X)=Cov(X,Y). 2. Scale-Dependent, i.e. when the scales of Y or X change, so is their covariance. Let Y1=a+bY,X1=c+dX. Then we have 3. Take values from - to + since b and d can take any values. Thus Cov(Y,X) does not measure the strength of the linear relationship between Y and X. 11/10/2018 ST3131, Lecture 2

Correlation Coefficient between Y and X Correlation Coefficient of Y and X is defined as the covariance of Standardized Y and X i.e. the covariance of Y and X divided by their standard deviations. Clearly Cor(Y,X) and Cov(Y,X) have the same signs so that it is also an indicator of the direction of the linear relationship between Y and X 1) Positive when Cor(Y,X)>0 2) Negative when Cor(Y,X)<0 3) Linear Uncorrelated when Corr(Y,X)=0 11/10/2018 ST3131, Lecture 2

Properties of Cor(Y,X) Symmetric, I.e. Cor(Y,X)=Cor(X,Y). Scale-Invariant, i.e., Not change with change of the scales of Y and X. Let Y1=a+bY,X1=c+dX, b>0,d>0. Then 3. Take values between –1 and 1. The strength of the Linear Relationship: 1). Strong when |Cor(Y,X)| close to 1; 2). Weak when |Cor(Y,X)| close to 0; 3). Linear Uncorrelated when Cor(Y,X)=0; But Y and X can still have some relationship. Counter example: the top-right picture where Y=2-cos(6.28X) (perfect nonlinear relationship) while Cor(Y,X)=0. 11/10/2018 ST3131, Lecture 2

Examples of Correlation Coefficients Cor(Y,X)=.98 Very Strong Linearity Cor(Y,X)=.71 Strong Linearity Cor(Y,X)=-.09 Near Uncorrelated Robustness: Both Cov(Y,X) and Cor(Y,X) are NOT Robust Statistics since their values will be affected by a few outliers . Examples: Anscombe quartets (see next slide) have the same summary statistics but quite different pictures: (a). can be described by a linear model; (b). can be described by a qudratic model; (c ). has an outlier, and so is (d). 11/10/2018 ST3131, Lecture 2

Anscombe Quartets (b) Strong Nonlinearity (a) Strong linearity (d) An outlier appears 11/10/2018 (c) an outlier appears ST3131, Lecture 2

1 2 … n i …. Table for Computing Variance and Covariance Total Note that Var(Y)=sum of squared deviations of Y divided by (n-1) Var(X)=sum of squared deviations of X divided by (n-1) Cov(Y,X)=sum of products of deviations of Y and X divided by (n-1). 11/10/2018 ST3131, Lecture 2

Example Computer Repair Data (Table 2.5, Page 27, see Table 2.6, Page 28 for detail computation) Conclusion: Y and X are strongly linearly related. Drawback: Cor(Y,X) can not be used to predict Y values given X values. This can be done with Simple Linear Regression Analysis. 11/10/2018 ST3131, Lecture 2

Strict Derivation of the LS-estimators SLR Model Intercept =the predicted value of Y when X=0, Slope =the change in Y for unit change in X. Least Squares Method: Find to minimize the Sum of the Squared Errors (SSE): The minimizers are: the same as those in Slide 7. 11/10/2018 ST3131, Lecture 2

Proof 11/10/2018 ST3131, Lecture 2

Proof (continued) Equality holds when Which are the least squares estimators of the parameters of and Since , we have an important property of Cor(Y,X) Moreover, we have another important equation: 11/10/2018 ST3131, Lecture 2

Computer Repair Data(continued), we have Example Computer Repair Data(continued), we have , Cov(Y,X)=136, Var(X)=2.96 Thus the LS- Regression Line is The fitted values and residuals then are In other words, we have : Minutes =4.162+15.5*Units Using this formula, we can compute the fitted (predicted) values, e.g. X=4, fitted value=4.162+15.5*4=66.20. X=11, predicted value=4.162+15.5*11=174.66. 11/10/2018 ST3131, Lecture 2

Exercise (1) Fill the following table, then compute the mean, variance, std of Y and X (2) Compute the covariance and Correlation of Y and X. (3) Compute the Simple Linear Regression Coefficients i 1 -.3 .09 .1 -.9 .81 .27 2 -.2 .04 .4 -.6 .36 .12 3 -.1 .01 .7 4 1.2 .2 5 1.6 .6 6 .3 2.0 Total 6.0 Statistics 1.0 11/10/2018 ST3131, Lecture 2

Review Sections 2.1-2.5 of Chapter 2. Reading Assignment Review Sections 2.1-2.5 of Chapter 2. Read Sections 2.6-2.9 of Chapter 2. Consider problems: a) How to do significance tests of parameters? b) How to construct confidence intervals of parameters? c) How to do inferences about prediction? 11/10/2018 ST3131, Lecture 2