Making a curved line straight Data Transformation & Regression.

Slides:



Advertisements
Similar presentations
Simple linear models Straight line is simplest case, but key is that parameters appear linearly in the model Needs estimates of the model parameters (slope.
Advertisements

Assumptions underlying regression analysis
Chapter 12 Inference for Linear Regression
Lesson 10: Linear Regression and Correlation
Kin 304 Regression Linear Regression Least Sum of Squares
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Forecasting Using the Simple Linear Regression Model and Correlation
Chapter 10 Curve Fitting and Regression Analysis
Université d’Ottawa / University of Ottawa 2001 Bio 4118 Applied Biostatistics L10.1 CorrelationCorrelation The underlying principle of correlation analysis.
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Linear Regression and Correlation
Chapter Eighteen MEASURES OF ASSOCIATION
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Correlation. Two variables: Which test? X Y Contingency analysis t-test Logistic regression Correlation Regression.
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Simple Linear Regression Lecture for Statistics 509 November-December 2000.
In previous lecture, we dealt with the unboundedness problem of LPM using the logit model. In this lecture, we will consider another alternative, i.e.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Business Statistics - QBM117 Statistical inference for regression.
Simple Linear Regression and Correlation
Checking Regression Model Assumptions NBA 2013/14 Player Heights and Weights.
Inference for regression - Simple linear regression
Regression Maarten Buis Outline Recap Estimation Goodness of Fit Goodness of Fit versus Effect Size transformation of variables and effect.
1 © 2008 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 5 Summarizing Bivariate Data.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Correlation and Prediction Error The amount of prediction error is associated with the strength of the correlation between X and Y.
Summarizing Bivariate Data
Examining Relationships in Quantitative Research
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Simple Linear Regression ANOVA for regression (10.2)
1 Regression Analysis The contents in this chapter are from Chapters of the textbook. The cntry15.sav data will be used. The data collected 15 countries’
REGRESSION DIAGNOSTICS Fall 2013 Dec 12/13. WHY REGRESSION DIAGNOSTICS? The validity of a regression model is based on a set of assumptions. Violation.
Multivariate Data Analysis Chapter 2 – Examining Your Data
Chapter 2 Examining Relationships.  Response variable measures outcome of a study (dependent variable)  Explanatory variable explains or influences.
LESSON 6: REGRESSION 2/21/12 EDUC 502: Introduction to Statistics.
Linear Correlation (12.5) In the regression analysis that we have considered so far, we assume that x is a controlled independent variable and Y is an.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
4 basic analytical tasks in statistics: 1)Comparing scores across groups  look for differences in means 2)Cross-tabulating categoric variables  look.
Variance Stabilizing Transformations. Variance is Related to Mean Usual Assumption in ANOVA and Regression is that the variance of each observation is.
Method 3: Least squares regression. Another method for finding the equation of a straight line which is fitted to data is known as the method of least-squares.
Multivariate Transformation. Multivariate Transformations  Started in statistics of psychology and sociology.  Also called multivariate analyses and.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
AP Statistics Review Day 1 Chapters 1-4. AP Exam Exploring Data accounts for 20%-30% of the material covered on the AP Exam. “Exploratory analysis of.
Chapter 12: Correlation and Linear Regression 1.
Inference about the slope parameter and correlation
Lecture 9 Sections 3.3 Objectives:
LEAST – SQUARES REGRESSION
Model validation and prediction
Kin 304 Regression Linear Regression Least Sum of Squares
Inference for Regression
Determination of toxic doses from log normal and log logistic models
BPK 304W Regression Linear Regression Least Sum of Squares
Checking Regression Model Assumptions
More Regression.
Simple Linear Regression
Checking Regression Model Assumptions
The Gamma PDF Eliason (1993).
The greatest blessing in life is
Chapter 14 Inference for Regression
Review Homework.
Adequacy of Linear Regression Models
Adequacy of Linear Regression Models
Checking the data and assumptions before the final analysis.
Adequacy of Linear Regression Models
Adequacy of Linear Regression Models
Regression and Correlation of Data
Inference for Regression
Presentation transcript:

Making a curved line straight Data Transformation & Regression

Last Class  Predicting the dependant variable and standard errors of predicted values.  Outliers.  Need to visually inspect data in graphic form.  Making a curved line straight.  Transformation.

Early Growth Pattern of Plants

y = Ln(y)

Early Growth Pattern of Plants y =  y

Homogeneity of Error Variance

y =Ln(y)

Growth Curve Y = e x

Growth Curve Y = Log(x)

Sigmoid Growth Curve 

 Accululative Normal Distribution

Sigmoid Growth Curve  Accululative Normal Distribution  T T-- T T-- ƒ (  d  d  T

Sigmoid Growth Curve  Accululative Normal Distribution  T T-- T T-- ƒ (  d  d  T

Probit Analysis Group of plants/insects exposed to different concentrations of a specific stimulant (i.e. insecticide). Group of plants/insects exposed to different concentrations of a specific stimulant (i.e. insecticide). Data are counts (or proportions), say number killed. Data are counts (or proportions), say number killed. Usually concerned or interested in concentration which causes specific event (i.e. LD 50%). Usually concerned or interested in concentration which causes specific event (i.e. LD 50%).

Probit Analysis ~ Example

Estimating the Mean  y = 50% Killed  x ~ 2.8

Estimating the Standard Deviation 2.8

2.8 2222

2.8 2222 95% values Estimating the Standard Deviation

2.8 2222 95% values  = 1.2 Estimating the Standard Deviation

Probit Analysis

Probit (  ) =  + . Log 10 (concentration)  =  = Log 10 (conc) to kill 50% (LD-50) is probit 0.5 = 0 0 = x LD-50 LD-50 = = 2.65%

Problems  Obtaining “good estimates” of the mean and standard deviation of the data.  Make a calculated guess, use iteration to get “better fit” to observed data.

Where Straight Lines Meet

Optimal Assent

Y 1 =a 1 +b 1 x

Optimal Assent Y 1 =a 1 +b 1 x Y 2 =a 2 +b 2 x

Optimal Assent Y 1 =a 1 +b 1 x Y 2 =a 2 +b 2 x t =[b 1 -b 2 ]/se(b) = ns = ns

Optimal Assent Y 1 =a 1 +b 1 x Y 3 =a 3 +b 3 x

Optimal Assent Y 1 =a 1 +b 1 x Y 3 =a 3 +b 3 x t =[b 1 -b 3 ]/se(b) = *** = ***

Optimal Assent Y 1 =a 1 +b 1 x Y 3 =a 3 +b 3 x

Optimal Assent Y 1 =a 1 +b 1 x Y 3 =a 3 +b 3 x t =[b 1 -b n ]/se(b) = *** = *** Y n =a n +b n x

Optimal Assent Y 1 =a 1 +b 1 x Y 3 =a 3 +b 3 x

Yield and Nitrogen

What application of nitrogen will result in the optimum yield response?

Intersecting Lines

Y = 2.81x Y = 9.01x

Intersecting Lines t = [b 11 - b 21 ]/average se(b) 6.2/0.593 = *, With 3 df Intersect = same value of y b 10 + b 11 x = y = b 20 + b 21 x x = [b 20 - b 10 ]/[b 11 - b 21 ] = lb N/acre with lb/acre seed yield

Intersecting Lines Y = 2.81x Y = 9.01x lb N/acre lb/acre

Linear Y = b 0 + b 1 x Quadratic Y = b 0 + b 1 x + b 2 x 2 Cubic Y = b 0 + b 1 x + b 2 x 2 + b 3 x 3 Bi-variate Distribution Correlation