Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

The Simple Regression Model
Chapter 12 Simple Linear Regression
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Correlation and regression Dr. Ghada Abo-Zaid
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
Ch11 Curve Fitting Dr. Deshi Ye
The General Linear Model. The Simple Linear Model Linear Regression.
The Use and Interpretation of the Constant Term
LINEAR REGRESSION MODEL
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
The Simple Linear Regression Model: Specification and Estimation
Linear Regression.
Chapter 10 Simple Regression.
Chapter 12 Simple Regression
Chapter 4 Multiple Regression.
Chapter 9 Simultaneous Equations Models. What is in this Chapter? In Chapter 4 we mentioned that one of the assumptions in the basic regression model.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
SIMPLE LINEAR REGRESSION
Chapter 11 Multiple Regression.
SIMPLE LINEAR REGRESSION
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Simple Linear Regression and Correlation
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
Regression and Correlation Methods Judy Zhong Ph.D.
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
3.1 Ch. 3 Simple Linear Regression 1.To estimate relationships among economic variables, such as y = f(x) or c = f(i) 2.To test hypotheses about these.
Chapter 11 Simple Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Simultaneous Equations Models (聯立方程式模型)
CHAPTER 14 MULTIPLE REGRESSION
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
MARKETING RESEARCH CHAPTER 18 :Correlation and Regression.
Prediction, Goodness-of-Fit, and Modeling Issues Prepared by Vera Tabakova, East Carolina University.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
1 B IVARIATE AND MULTIPLE REGRESSION Estratto dal Cap. 8 di: “Statistics for Marketing and Consumer Research”, M. Mazzocchi, ed. SAGE, LEZIONI IN.
Multiple Regression I 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 1) Terry Dielman.
Correlation & Regression Analysis
Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and Alison Kelly Copyright © 2014 by McGraw-Hill Higher Education. All rights.
Introduction. We want to see if there is any relationship between the results on exams and the amount of hours used for studies. Person ABCDEFGHIJ Hours/
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
The simple linear regression model and parameter estimation
Chapter 4 Basic Estimation Techniques
Chapter 7. Classification and Prediction
The Simple Linear Regression Model: Specification and Estimation
Chapter 5: The Simple Regression Model
Chapter 5 STATISTICS (PART 4).
Undergraduated Econometrics
The Simple Linear Regression Model: Specification and Estimation
Simple Linear Regression
SIMPLE LINEAR REGRESSION
Product moment correlation
SIMPLE LINEAR REGRESSION
Linear Regression Summer School IFPRI
Introduction to Regression
Presentation transcript:

Chapter 3 Simple Regression

What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the assumptions of this basic model It then discusses two methods of estimation: the method of moments (MM) and the method of least squares (LS). The method of maximum likelihood (ML) is discussed in the appendix

3.1 Introduction Example 1: Simple Regression y = sale x = advertising expenditures Here we try to determine the relationship between sales and advertising expenditures.

3.1 Introduction Example 2: Multiple Regression y = consumption expenditures of a family x 1 = family income x 2 = financial assets of the family x 3 = family size

3.1 Introduction There are several objectives in studying these relationships. They can be used to: 1. Analyze the effects of policies that involve changing the individual x's. In Example 1 this involves analyzing the effect of changing advertising expenditures on sales 2. Forecast the value of y for a given set of x's. 3. Examine whether any of the x's have a significant effect on y.

3.1 Introduction Given the way we have set up the problem until now, the variable y and the x variables are not on the same footing Implicitly we have assumed that the x's are variables that influence y or are variables that we can control or change and y is the effect variable. There are several alternative terms used in the literature for y and x 1, x 2,..., x k. These are shown in Table 3.1.

3.1 Introduction Table 3.1 Classification of variables in regression analysis (a)Presdictand Predictors (b)Regressand Regressors (c)Explained variable Explanatory variables (d)Dependent variable Independent variables (e)Effect variable Causal variables (f)Endogenous variable Exogenous variables (g)Target variable Control variables y x 1, x 2,……, x k

3.2 Specification of the Relationships As mentioned in Section 3.1, we will discuss the case of one explained (dependent) variable, which we denote by y, and one explanatory (independent) variable, which we denote by x. The relationship between y and x is denoted by y = f(x) Where f(x) is a function of x

3.2 Specification of the Relationships Going back to equation (3.1), we will assume that the function f(x) is linear in x, that is, And we will assume that this relationship is a stochastic relationship, that is, Where,which is called an error or disturbance, has a known probability distribution (i.e., is a random variable).

3.2 Specification of the Relationships

In equation (3.2), is the deterministic component of y and u is the stochastic or random component. and are called regression coefficients or regression parameters that we estimate from the data on y and x

3.2 Specification of the Relationships Why should we add an error term u ? What are the sources of the error term u in equation (3.2)? There are three main sources :

3.2 Specification of the Relationships 1.Unpredictable element of randomness in human responses. ex. If y =consumption expenditure of a household and x = disposable income of the household, there is an unpredictable element of randomness in each household's consumption. The household does not behave like a machine.

3.2 Specification of the Relationships 2.Effect of a large number of omitted variables. Again in our example x is not the only variable influencing y. The family size, tastes of the family, spending habits, and so on, affect the variable y. The error u is a catchall for the effects of all these variables, some of which may not even be quantifiable, and some of which may not even be identifiable.

3.2 Specification of the Relationships 3. Measurement error in y. In our example this refers to measurement error in the household consumption. That is, we cannot measure it accurately.

3.2 Specification of the Relationships If we have n observations on y and x, we can write equation (3.2) as Our objective is to get estimates of the unknown parameters and in equation (3.3) given the n observations on y and x.

3.2 Specification of the Relationships To do this we have to make some assumption about the error terms. The assumptions we make are: 1.Zero mean. 2.Common variance. 3.Independence. and are independent for all

3.2 Specification of the Relationships 4.Independence of. and are independent for all i and j. This assumption automatically follows if are considered nonrandom variables. With reference to Figure 3.1, what this says is that the distribution of u does not depend on the value of x. 5.Normality, are normally distributed for all i. In conjunction with assumptions 1, 2, and 3 this implies that are independently and normally distributed with mean zero and a common variance. We write this as

3.2 Specification of the Relationships These are the assumptions with which we start. We will, however, relax some of these assumptions in later chapters. Assumption 2 is relaxed in Chapter 5. Assumption 3 is relaxed in Chapter 6. Assumption 4 is relaxed in Chapter 9.

3.2 Specification of the Relationships We will discuss three methods for estimating the parameters and : 1. The method of moments (MM). 2. The method of least squares (LS). 3. The method of maximum likelihood (ML).

3.3 The Method of Moments The assumptions we have made about the error term u imply that In the method of moments, we replace these conditions by their sample counterparts. Let and be the estimators for and, respectively. The sample counterpart of is the estimated error (which is also called the residual), defined as

3.3 The Method of Moments The two equations to determine and are obtained by replacing population assumptions by their sample counterparts:

3.3 The Method of Moments In these and the following equations, denotes. Thus we get the two equations These equations can be written as (noting that )

3.4 The Method of Least Squares The method of least squares requires that we should choose and as estimates of and, respectively, so that is a minimum. Q is also the sum of squares of the (within- sample) prediction errors when we predict given and the estimated regression equation. We will show in the appendix to this chapter that the least squares estimators have desirable optimal properties.

3.4 The Method of Least Squares or (3.6) and (3.7)

3.4 The Method of Least Squares Let us define and

3.4 The Method of Least Squares The residual sum of squares (to be denoted by RSS) is given by

3.4 The Method of Least Squares But.Hence we have is usually denoted by TSS (total sum of squares) and is usually denoted by ESS (explained sum of squares). Thus TSS = ESS + RSS ( total) (explained) (residual)

3.4 The Method of Least Squares The proportion of the total sum of squares explained is denoted by,where is called the correlation coefficient. Thus and.If is high (close to 1), then x is a good “explanatory” variable for y. The term is called the correlation determination and must fall between zero and 1 for any given regression.

3.4 The Method of Least Squares If is close to zero, the variable x explains very little of the variation in y. If is close to 1, the variable x explains most of the variation in y. The coefficient of determination is given by

3.9 Alternative Functional Forms for Regression Equations For instance, for the data points depicted in Figure 3.7(a), where y is increasing more slowly than x, a possible functional form is y = α +β logx. This is called a semi-log form, since it involves the logarithm of only one of the two variables x and y.

3.9 Alternative Functional Forms for Regression Equations In this case, if we redefine a variable X = log x, the equation becomes y = α + βX. Thus we have a linear regression model with the explained variable y and the explanatory variable X = log x.

3.9 Alternative Functional Forms for Regression Equations For the data points depicted in Figure 3.7(b), where y is increasing faster than x, a possible functional from is. In this case we take logs of both sides and get another kind of semi-log specification: If we define Y= log y and, we have which is in the form of a linear regression equation.

3.9 Alternative Functional Forms for Regression Equations

An alternative model one can use is In this case taking logs of both sides, we get In this case can be interpreted as an elasticity. Hence this form is popular in econometric work. This is call a double-log specification since it involves logarithms of both x and y. Now define Y= log y, X= log x, and.We have which is in the form of a linear regression equation. An illustrative example is given at the end of this section.

3.9 Alternative Functional Forms for Regression Equations Some other functional forms that are useful when the data points are as shown in Figure 3.8 are or In the first case we define X=1/x and in the second case we define.In both case the equation is linear in the variables after the transformation.

3.9 Alternative Functional Forms for Regression Equations Some other nonlinearities can be handled by what is known as “search procedures.” For instance, suppose that we have the regress equation The estimates of,,and are obtained by minimizing