Some terminology When the relation between variables are expressed in this manner, we call the relevant equation(s) mathematical models The intercept and.

Slides:



Advertisements
Similar presentations
Things to do in Lecture 1 Outline basic concepts of causality
Advertisements

Regression and correlation methods
The Simple Regression Model
Chapter 7 Statistical Data Treatment and Evaluation
1 MA 1128: Lecture 19 – 4/19/11 Quadratic Formula Solving Equations with Graphs.
1 Functions and Applications
Chapter 8 Linear Regression.
 Coefficient of Determination Section 4.3 Alan Craig
Quantitative Methods 2 Lecture 3 The Simple Linear Regression Model Edmund Malesky, Ph.D., UCSD.
Regression What is regression to the mean?
Copyright © 2009 Pearson Education, Inc. Chapter 8 Linear Regression.
CHAPTER 8: LINEAR REGRESSION
Turning Point At the beginning of the course, we discussed three ways in which mathematics and statistics can be used to facilitate psychological science.
Understanding the General Linear Model
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Statistics for the Social Sciences
Chapter 4 Multiple Regression.
The Basics of Regression continued
Nonlinear Regression Probability and Statistics Boris Gervits.
Chapter Sampling Distributions and Hypothesis Testing.
Lecture 5 Curve fitting by iterative approaches MARINE QB III MARINE QB III Modelling Aquatic Rates In Natural Ecosystems BIOL471 © 2001 School of Biological.
Analysis of Individual Variables Descriptive – –Measures of Central Tendency Mean – Average score of distribution (1 st moment) Median – Middle score (50.
CHAPTER 3 Describing Relationships
Math 34A Chapter 5 examples Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB.
PSY 307 – Statistics for the Behavioral Sciences Chapter 7 – Regression.
Correlation and Regression
Lecture 3: Bivariate Data & Linear Regression 1.Introduction 2.Bivariate Data 3.Linear Analysis of Data a)Freehand Linear Fit b)Least Squares Fit c)Interpolation/Extrapolation.
EC339: Lecture 6 Chapter 5: Interpreting OLS Regression.
Testing Theories: Three Reasons Why Data Might not Match the Theory Psych 437.
Basic linear regression and multiple regression Psych Fraley.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Linear Functions 2 Sociology 5811 Lecture 18 Copyright © 2004 by Evan Schofer Do not copy or distribute without permission.
Least-Squares Regression Section 3.3. Why Create a Model? There are two reasons to create a mathematical model for a set of bivariate data. To predict.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Chapter 20 Linear Regression. What if… We believe that an important relation between two measures exists? For example, we ask 5 people about their salary.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 8 Linear Regression.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Relationships If we are doing a study which involves more than one variable, how can we tell if there is a relationship between two (or more) of the.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Regression Understanding relationships and predicting outcomes.
CHAPTER 8 Linear Regression. Residuals Slide  The model won’t be perfect, regardless of the line we draw.  Some points will be above the line.
Correlation & Regression Analysis
Regression Analysis: Part 2 Inference Dummies / Interactions Multicollinearity / Heteroscedasticity Residual Analysis / Outliers.
Chapter 10: Determining How Costs Behave 1 Horngren 13e.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
Residuals Recall that the vertical distances from the points to the least-squares regression line are as small as possible.  Because those vertical distances.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide 8- 1.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Curve Fitting Pertemuan 10 Matakuliah: S0262-Analisis Numerik Tahun: 2010.
Week 6. Statistics etc. GRS LX 865 Topics in Linguistics.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
1 HETEROSCEDASTICITY: WEIGHTED AND LOGARITHMIC REGRESSIONS This sequence presents two methods for dealing with the problem of heteroscedasticity. We will.
Copyright © Cengage Learning. All rights reserved. Sequences and Series.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
CHAPTER- 3.2 ERROR ANALYSIS. 3.3 SPECIFIC ERROR FORMULAS  The expressions of Equations (3.13) and (3.14) were derived for the general relationship of.
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar 8: Time Series Analysis & Forecasting – Part 1
Copyright © Cengage Learning. All rights reserved. 8 9 Correlation and Regression.
1 Functions and Applications
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
Simple Linear Regression
Chapter 8 Part 2 Linear Regression
MBF1413 | Quantitative Methods Prepared by Dr Khairul Anuar
Least-Squares Regression
Least-Squares Regression
Product moment correlation
Linear Regression Dr. Richard Jackson
Presentation transcript:

Some terminology When the relation between variables are expressed in this manner, we call the relevant equation(s) mathematical models The intercept and weight values are called the parameters of the model. Although one can describe the relationship between two variables in the way we have done here, for now on we’ll assume that our models are causal models, such that the variable on the left-hand side of the equation is being caused by the variable(s) on the right side.

Terminology The values of Y in these models are often called predicted values, sometimes abbreviated as Y-hat or. Why? They are the values of Y that are implied or predicted by the specific parameters of the model.

Parameter Estimation Up to this point, we have assumed that our basic models are correct. There are two important issues we need to deal with, however: –Is the basic model correct (regardless of the value of the parameters)? That is, is a linear, as opposed to a quadratic, model the appropriate model for characterizing the relationship between variables? –If the model is correct, what are the most right parameter values for the model?

Parameter Estimation For now, we will continue to assume that the basic model is correct. In the third part of the course, we will deal with methods for comparing alternative models. The process of obtaining the correct parameter values (assuming we are working with the right model) is called parameter estimation.

Parameter Estimation Often, theories specify the form of the relationship rather than the specific values of the parameters The parameters themselves, assuming the basic model is correct, are typically estimated from data. We refer to the estimation processes as “calibrating the model.” We need a method for choosing parameter values that will give us the best representation of the data possible.

Parameter Estimation example Let’s assume that we believe there is a linear relationship between X and Y. Assume we have collected the following data Which set of parameter values will bring us closest to representing the data accurately?

Estimation example We begin by picking some values, plugging them into the equation, and seeing how well the implied values correspond to the observed values We can quantify what we mean by “how well” by examining the difference between the model-implied Y and the actual Y value this difference,, is often called error in prediction

Estimation example Let’s try a different value of b and see what happens Now the implied values of Y are getting closer to the actual values of Y, but we’re still off by quite a bit

Estimation example Things are getting better, but certainly things could improve

Estimation example Ah, much better

Estimation example Now that’s very nice There is a perfect correspondence between the implied values of Y and the actual values of Y

Estimation example Whoa. That’s a little worse. Simply increasing b doesn’t seem to make things increasingly better

Estimation example Ugg. Things are getting worse again.

Parameter Estimation example Here is one way to think about what we’re doing: –We are trying to find a set of parameter values that will give us a small—the smallest—discrepancy between the predicted Y values and the actual values of Y. How can we quantify this?

Parameter Estimation example One way to do so is to find the difference between each value of Y and the corresponding predicted value (we called these differences “errors” before), square these differences, and average them together

Parameter Estimation example The form of this equation should be familiar. Notice that it represents some kind of average of squared deviations This average is often called error variance. Sometimes people simply sum the squared errors. When expressed this way, the sum is often called the sum of squared errors or SSE.

Parameter Estimation example In estimating the parameters of our model, we are trying to find a set of parameters that minimizes the error variance. In other words, we want to be as small as it possibly can be. The process of finding this minimum value is called least- squares estimation.

Parameter Estimation example In this graph I have plotted the error variance as a function of the different parameter values we chose for b. Notice that our error was large at first, but got smaller as we made X larger. Eventually, the error reached a minimum and, then, began to increase again as we made X larger.

Parameter Estimation example The minimum in this example occurred when b = 2. This is the “best” value of b, when we define “best” as the value that minimizes the error variance. There is no other value of b that will make the error smaller. (0 is as low as you can go.)

Ways to estimate parameters The method we just used is sometimes called the brute force or gradient descent method to estimating parameters. –More formally, gradient decent involves starting with viable parameter value, calculating the error using slightly different value, moving the best guess parameter value in the direction of the smallest error, then repeating this process until the error is as small as it can be. Analytic methods –With simple linear models, the equation is so simple that brute force methods are unnecessary.

Analytic least-squares estimation Specifically, one can use calculus to find the values of a and b that will minimize the error function

Analytic least-squares estimation When this is done (we won’t actually do the calculus here ), the obtain the following equations:

Analytic least-squares estimation Thus, we can easily find the least-squares estimates of a and b from simple knowledge of (1) the correlation between X and Y, (2) the SD’s of X and Y, and (3) the means of X and Y:

A neat fact Notice what happens when X and Y are in standard score form Thus,

In this past example, we have dealt with a situation in which a linear model of the form Y = 2 + 2X perfectly accounts for the data. (That is, there is no discrepancy between the values implied by the model and the actual data.) When this is not true, we can still find least squares estimates of the parameters.