Non-linear relationships

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Analytic Comparisons & Trend Analyses Analytic Comparisons –Simple comparisons –Complex comparisons –Trend Analyses Errors & Confusions when interpreting.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
MANE 4240 & CIVL 4240 Introduction to Finite Elements Numerical Integration in 1D Prof. Suvranu De.
(Correlation and) (Multiple) Regression Friday 5 th March (and Logistic Regression too!)
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Classification and Prediction: Regression Analysis
Relationships Among Variables
Objectives of Multiple Regression
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
One-Factor Experiments Andy Wang CIS 5930 Computer Systems Performance Analysis.
1 FORECASTING Regression Analysis Aslı Sencer Graduate Program in Business Information Systems.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Regression and correlation analysis (RaKA) 1. Investigating the relationships between the statistical characteristics: 2 Investigating the relationship.
Curve-Fitting Regression
General Linear Models; Generalized Linear Models Hal Whitehead BIOL4062/5062.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Regression and correlation Dependence of two quantitative variables.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
CORRELATION-REGULATION ANALYSIS Томский политехнический университет.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Computational Intelligence: Methods and Applications Lecture 14 Bias-variance tradeoff – model selection. Włodzisław Duch Dept. of Informatics, UMK Google:
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard)   Week 5 Multiple Regression  
The simple linear regression model and parameter estimation
Chapter 4: Basic Estimation Techniques
Chapter 7. Classification and Prediction
Regression Analysis AGEC 784.
Statistical Data Analysis - Lecture /04/03
Correlation I have two variables, practically „equal“ (traditionally marked as X and Y) – I ask, if they are independent and if they are „correlated“,
Workshop on Residential Property Price Indices
3.1 Examples of Demand Functions
The Correlation Coefficient (r)
Discrete distributions and their use
Kin 304 Regression Linear Regression Least Sum of Squares
CH 5: Multivariate Methods
Essentials of Modern Business Statistics (7e)
Introduction to the General Linear Model (GLM)
Checking Regression Model Assumptions
Physics 114: Exam 2 Review Material from Weeks 7-11
Regression Analysis Week 4.
Statistical Methods For Engineers
CHAPTER 29: Multiple Regression*
Checking Regression Model Assumptions
Regression Models - Introduction
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Hypothesis testing and Estimation
Linear regression Fitting a straight line to observations.
Least Squares Fitting A mathematical procedure for finding the best-fitting curve to a given set of points by minimizing the sum of the squares of the.
Nonlinear Fitting.
Product moment correlation
SIMPLE LINEAR REGRESSION
Diagnostics and Remedial Measures
One-Factor Experiments
Introduction to Regression
3.2. SIMPLE LINEAR REGRESSION
Generalized Additive Model
Diagnostics and Remedial Measures
MGS 3100 Business Analysis Regression Feb 18, 2016
Regression Models - Introduction
Statistical inference for the slope and intercept in SLR
The Correlation Coefficient (r)
Presentation transcript:

Non-linear relationships What to do, when the relationship is not linear

Possibilities Transformation – can practically help just in monotonous relationships – one have to be careful – transformation of predictor changes just shape, transformation of response changes also the probability characteristics (distribution of residuals)

Possibilities Polynomial regression – any function can be replaced (in limited range of predictor values) by polynomial function, i.e. b0 + b1x + b2x2 + b3x3... Assumption, residuals are normal, homogeneously distributed, i.e Y = b0 + b1x + b2x2 + b3x3…+ε Traditional names quadratic regression, cubic regression

Polynomial regression Actually, it is application of multiple linear regression, where predictors are X, X2, X3 etc. Computation is the same (i.e. again criteria of least sum of residual squares, which again has (in normal conditions) one minimum). Similar meaning has also R2, similar computation of significance tests (i.e. total ANOVA of model, and tests for single terms of polynomial). So, I assume again, that ε is additive, independent of predicted value (homogeneity of variance).

With degree of polynomial increases “flexibility” 1 2 Time, in hours Time, in hours Weight, in kilograms Weight, in kilograms 3 4 Time, in hours Time, in hours Weight, in kilograms Weight, in kilograms 5 Attention! Increasing complexity does not always mean better prediction ability. Time, in hours Weight, in kilograms

Stepwise regression – it makes model more complex gradually

Stepwise regression – it makes model more complex gradually -quadratic regression can be highly significant even if linear regression is not Significance of quadratic term can be understood as prove of relation non-linearity

We usually use polynomial regression, when we see the relation isn’t linear, but we haven’t any idea, what shape does the function have I do not remember seeing wise use of higher than third order polynomial.

Other possibilities I have idea (e.g. from some theory), how should dependence looks like and I believe the residuals will be randomly spread around value predicted, i.e. model is Y=f(X) + ε [X here is vector, so, it can be more then one independent variable] We estimate again using method of least sum of squares

In contrast to methods of linear regression (including polynomial one) it is necessary to find minimum using methods of numerical mathematics – there haven’t to exist analytical solution, nor there is any certainty, that minimum found is global one. Numerical progress: 1. Derivate according to all parameters estimated. 2. Take all the derivations as equal to zero. 3. Solve the system. Numerical solution of formula f(x)=0

In contrast to methods of linear regression (including polynomial one) it is necessary to find minimum using methods of numerical mathematics – there haven’t to exist analytical solution, nor there is any certainty, that minimum found is global one. Numerical progress: 1. Derivate according to all parameters estimated. 2. Take all the derivations as equal to zero. 3. Resolve the system. Numerical solution of formula f(x)=0 [Newton method] f(x) x1 x2 x3 x “My” estimation of x

Disadvantages of numerical solution It doesn’t always converge It sometimes finds just local minimum (derivations are equal to zero even there), and we haven’t many possibilities how to prove, which minimum it is. We need initial values of parameters.

Analogy – taw is falling down

Various “local regressions” – I won't get function, it is different for various local parts of the line

I know distribution of response variable Generalized Linear Models They are able to reflect distribution type - so, even which values can the response have (e.g. probability of survival must be between zero and one) Link function

Typical example - logistic regression