Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Linear Regression.
Data Modeling and Parameter Estimation Nov 9, 2005 PSCI 702.
Experimental Design, Response Surface Analysis, and Optimization
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Visual Recognition Tutorial
Jonathan Richard Shewchuk Reading Group Presention By David Cline
PHYS2020 NUMERICAL ALGORITHM NOTES ROOTS OF EQUATIONS.
Function Optimization Newton’s Method. Conjugate Gradients
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Chapter 1 Introduction The solutions of engineering problems can be obtained using analytical methods or numerical methods. Analytical differentiation.
Curve-Fitting Regression
SYSTEMS Identification
Development of Empirical Models From Process Data
Nonlinear Regression Probability and Statistics Boris Gervits.
Prediction and model selection
Engineering Computation Curve Fitting 1 Curve Fitting By Least-Squares Regression and Spline Interpolation Part 7.
Systems of Non-Linear Equations
Chi Square Distribution (c2) and Least Squares Fitting
Optimization Methods One-Dimensional Unconstrained Optimization
Simple Linear Regression Analysis
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: nonlinear regression Original citation: Dougherty, C. (2012) EC220 -
Collaborative Filtering Matrix Factorization Approach
Neural Networks Lecture 8: Two simple learning algorithms
Objectives of Multiple Regression
Least-Squares Regression
Ch 8.1 Numerical Methods: The Euler or Tangent Line Method
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
3/2003 Rev 1 I – slide 1 of 33 Session I Part I Review of Fundamentals Module 2Basic Physics and Mathematics Used in Radiation Protection.
Lecture Notes Dr. Rakhmad Arief Siregar Universiti Malaysia Perlis
Analytical vs. Numerical Minimization Each experimental data point, l, has an error, ε l, associated with it ‣ Difference between the experimentally measured.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 4 Curve Fitting.
Applications The General Linear Model. Transformations.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
MECN 3500 Inter - Bayamon Lecture 9 Numerical Methods for Engineering MECN 3500 Professor: Dr. Omar E. Meza Castillo
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Curve-Fitting Regression
Response surfaces. We have a dependent variable y, independent variables x 1, x 2,...,x p The general form of the model y = f(x 1, x 2,...,x p ) +  Surface.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
1 NONLINEAR REGRESSION Suppose you believe that a variable Y depends on a variable X according to the relationship shown and you wish to obtain estimates.
Boyce/DiPrima 9 th ed, Ch 11.3: Non- Homogeneous Boundary Value Problems Elementary Differential Equations and Boundary Value Problems, 9 th edition, by.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Introduction to Statistical Quality Control, 4th Edition Chapter 13 Process Optimization with Designed Experiments.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Computacion Inteligente Least-Square Methods for System Identification.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
NUMERICAL ANALYSIS I. Introduction Numerical analysis is concerned with the process by which mathematical problems are solved by the operations.
The simple linear regression model and parameter estimation
Chapter 7. Classification and Prediction
The Maximum Likelihood Method
2.1 Equations of Lines Write the point-slope and slope-intercept forms
Non-Linear Regression
Collaborative Filtering Matrix Factorization Approach
Regression Models - Introduction
Nonlinear regression.
5.2 Least-Squares Fit to a Straight Line
5.4 General Linear Least-Squares
Simple Linear Regression
Regression Models - Introduction
Presentation transcript:

Non-Linear Models

Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring  ) “rate of increase in Y” =

The Logistic Growth Model or (ignoring  ) “rate of increase in Y” = Equation:

The Gompertz Growth Model: or (ignoring  ) “rate of increase in Y” = Equation:

Non-Linear Regression

Introduction Previously we have fitted, by least squares, the General Linear model which were of the type: Y =  0 +  1 X 1 +  2 X  p X p + 

the above equation can represent a wide variety of relationships. there are many situations in which a model of this form is not appropriate and too simple to represent the true relationship between the dependent (or response) variable Y and the independent (or predictor) variables X 1, X 2,... and X p. When we are led to a model of nonlinear form, we would usually prefer to fit such a model whenever possible, rather than to fit an alternative, perhaps less realistic, linear model. Any model which is not of the form given above will be called a nonlinear model.

This model will generally be of the form: Y = f(X 1, X 2,..., X p |  1,  2,...,  q ) +  where the function (expression) f is known except for the q unknown parameters  1,  2,...,  q.

Least Squares in the Nonlinear Case

Suppose that we have collected data on the Y, –(y 1, y 2,...y n ) corresponding to n sets of values of the independent variables X 1, X 2,... and X p – (x 11, x 21,..., x p1 ), –(x 12, x 22,..., x p2 ), –... and – (x 12, x 22,..., x p2 ).

For a set of possible values  1,  2,...,  q of the parameters, a measure of how well these values fit the model described in equation * above is the residual sum of squares function where is the predicted value of the response variable y i from the values of the p independent variables x 1i, x 2i,..., x pi using the model in equation * and the values of the parameters  1,  2,...,  q.

The Least squares estimates of  1,  2,...,  q, are values which minimize S(  1,  2,...,  q ). It can be shown that the error terms are independent normally distributed with mean 0 and common variance  2 than the least squares estimates are also the maximum likelihood estimate of  1,  2,...,  q ).

To find the least squares estimate we need to determine when all the derivatives S(  1,  2,...,  q ) with respect to each parameter  1,  2,... and  q are equal to zero. This quite often leads to a set of equations in  1,  2,... and  q that are difficult to solve even with one parameter and a comparatively simple nonlinear model. When more parameters are involved and the model is more complicated, the solution of the normal equations can be extremely difficult to obtain, and iterative methods must be employed.

To compound the difficulties it may happen that multiple solutions exist, corresponding to multiple stationary values of the function S(  1,  2,...,  q ). When the model is linear, these equations form a set of linear equations in  1,  2,... and  q which can be solved for the least squares estimates. In addition the sum of squares function, S(  1,  2,...,  q ), is a quadratic function of the parameters and is constant on ellipsoidal surfaces centered at the least squares estimates.

Example We have collected data on two variables (Y, X) The model we will consider is Y =  1 e -X +  2 +  The data: (x 1,y 1 ) (x 2,y 2 ) (x 3,y 3 ),..., (x n,y n )

The predicted value of y i using the model is: The Residual Sum of Squares is: The least squares estimates of  1 and  2 satisfy

Techniques for Estimating the Parameters of a Nonlinear System In some nonlinear problems it is convenient to determine equations (the Normal Equations) for the least squares estimates, the values that minimize the sum of squares function, S(  1,  2,...,  q ). These equations are nonlinear and it is usually necessary to develop an iterative technique for solving them.

In addition to this approach there are several currently employed methods available for obtaining the parameter estimates by a routine computer calculation. We shall mention three of these: –1)Steepest descent, –2)Linearization, and –3)Marquardt's procedure.

In each case a iterative procedure is used to find the least squares estimators. That is an initial estimates,,for these values are determined. The procedure than finds successfully better estimates, that hopefully converge to the least squares estimates,

Steepest Descent The steepest descent method focuses on determining the values of  1,  2,...,  q that minimize the sum of squares function, S(  1,  2,...,  q ). The basic idea is to determine from an initial point, and the tangent plane to S(  1,  2,...,  q ) at this point, the vector along which the function S(  1,  2,...,  q ) will be decreasing at the fastest rate. The method of steepest descent than moves from this initial point along the direction of steepest descent until the value of S(  1,  2,...,  q ) stops decreasing.

It uses this point, as the next approximation to the value that minimizes S(  1,  2,...,  q ). The procedure than continues until the successive approximation arrive at a point where the sum of squares function, S(  1,  2,...,  q ) is minimized. At that point, the tangent plane to S(  1,  2,...,  q ) will be horizontal and there will be no direction of steepest descent.

It should be pointed out that the technique of steepest descent is a technique of great value in experimental work for finding stationary values of response surfaces. While, theoretically, the steepest descent method will converge, it may do so in practice with agonizing slowness after some rapid initial progress.

Slow convergence is particularly likely when the S(  1,  2,...,  q ) contours are attenuated and banana-shaped (as they often are in practice), and it happens when the path of steepest descent zigzags slowly up a narrow ridge, each iteration bringing only a slight reduction in S(  1,  2,...,  q ).

A further disadvantage of the steepest descent method is that it is not scale invariant. The indicated direction of movement changes if the scales of the variables are changed, unless all are changed by the same factor. The steepest descent method is, on the whole, slightly less favored than the linearization method (described later) but will work satisfactorily for many nonlinear problems, especially if modifications are made to the basic technique.

Steepest descent path Initial guess Steepest Descent

Linearization The linearization (or Taylor series) method uses the results of linear least squares in a succession of stages. Suppose the postulated model is of the form: Y = f(X 1, X 2,..., X p |  1,  2,...,  q ) +  Let be initial values for the parameters  1,  2,...,  q. These initial values may be intelligent guesses or preliminary estimates based on whatever information are available.

These initial values will, hopefully, be improved upon in the successive iterations to be described below. The linearization method approximates f(X 1, X 2,..., X p |  1,  2,...,  q ) with a linear function of  1,  2,...,  q using a Taylor series expansion of f(X 1, X 2,..., X p |  1,  2,...,  q ) about the point and curtailing the expansion at the first derivatives. The method then uses the results of linear least squares to find values, that provide the least squares fit of of this linear function to the data.

The procedure is then repeated again until the successive approximations converge to hopefully at the least squares estimates:

Contours of RSS for linear approximatin Initial guess 2 nd guess Linearization

Contours of RSS for linear approximatin Initial guess 2 nd guess 3 rd guess

Contours of RSS for linear approximatin Initial guess 2 nd guess 3 rd guess 4 th guess

The linearization procedure has the following possible drawbacks: 1.It may converge very slowly; that is, a very large number of iterations may be required before the solution stabilizes even though the sum of squares S(  1,  2, …,  p ) may decrease consistently as the number of iterations increase. This sort of behavior is not common but can occur. 2.It may oscillate widely, continually reversing direction, and often increasing, as well as decreasing the sum of squares. Nevertheless the solution may stabilize eventually. 3.It may not converge at all, and even diverge, so that the sum of squares increases iteration after iteration without bound.

The linearization method is, in general, a useful one and will successfully solve many nonlinear problems. Where it does not, consideration should be given to reparameterization of the model or to the use of Marquardt's procedure.

Next topic: Examination of ResidualsExamination of Residuals