Semi-mechanistic modelling in Nonlinear Regression: a case study by Katarina Domijan 1, Murray Jorgensen 2 and Jeff Reid 3 1 AgResearch Ruakura 2 University.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
458 Quantifying Uncertainty using Classical Methods (Likelihood Profile, Bootstrapping) Fish 458, Lecture 12.
Experimental Design, Response Surface Analysis, and Optimization
Prediction, Correlation, and Lack of Fit in Regression (§11. 4, 11
P M V Subbarao Professor Mechanical Engineering Department
A Short Introduction to Curve Fitting and Regression by Brad Morantz
EARS1160 – Numerical Methods notes by G. Houseman
Aspects of Conditional Simulation and estimation of hydraulic conductivity in coastal aquifers" Luit Jan Slooten.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Active Calibration of Cameras: Theory and Implementation Anup Basu Sung Huh CPSC 643 Individual Presentation II March 4 th,
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Chapter 10 Simple Regression.
Integrated Assessment of Sustainability =Economic analysis +Non-market analysis.
Lecture 4 MGMT © 2011 Houman Younessi Supply: Production What is the optimal level of output? How should we decide among alternative production.
Nonlinear Regression Probability and Statistics Boris Gervits.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
D Nagesh Kumar, IIScOptimization Methods: M1L4 1 Introduction and Basic Concepts Classical and Advanced Techniques for Optimization.
Statistics 200b. Chapter 5. Chapter 4: inference via likelihood now Chapter 5: applications to particular situations.
Maximum likelihood (ML)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: prediction Original citation: Dougherty, C. (2012) EC220 - Introduction.
1 PREDICTION In the previous sequence, we saw how to predict the price of a good or asset given the composition of its characteristics. In this sequence,
Normalised Least Mean-Square Adaptive Filtering
Calibration & Curve Fitting
Collaborative Filtering Matrix Factorization Approach
Objectives of Multiple Regression
3. Multiple Regression Analysis: Estimation -Although bivariate linear regressions are sometimes useful, they are often unrealistic -SLR.4, that all factors.
2015 AprilUNIVERSITY OF HAIFA, DEPARTMENT OF STATISTICS, SEMINAR FOR M.A 1 Hastie, Tibshirani and Friedman.The Elements of Statistical Learning (2nd edition,
Calibration Guidelines 1. Start simple, add complexity carefully 2. Use a broad range of information 3. Be well-posed & be comprehensive 4. Include diverse.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Model Inference and Averaging
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Andrew Thomson on Generalised Estimating Equations (and simulation studies)
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Quality of Curve Fitting P M V Subbarao Professor Mechanical Engineering Department Suitability of A Model to a Data Set…..
01/20151 EPI 5344: Survival Analysis in Epidemiology Maximum Likelihood Estimation: An Introduction March 10, 2015 Dr. N. Birkett, School of Epidemiology,
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Curve-Fitting Regression
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Three Frameworks for Statistical Analysis. Sample Design Forest, N=6 Field, N=4 Count ant nests per quadrat.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
WCRP Extremes Workshop Sept 2010 Detecting human influence on extreme daily temperature at regional scales Photo: F. Zwiers (Long-tailed Jaeger)
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Regression Analysis1. 2 INTRODUCTION TO EMPIRICAL MODELS LEAST SQUARES ESTIMATION OF THE PARAMETERS PROPERTIES OF THE LEAST SQUARES ESTIMATORS AND ESTIMATION.
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
Anders Nielsen Technical University of Denmark, DTU-Aqua Mark Maunder Inter-American Tropical Tuna Commission An Introduction.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.
Fundamentals of Data Analysis Lecture 10 Correlation and regression.
Stats Methods at IC Lecture 3: Regression.
ECE3340 Numerical Fitting, Interpolation and Approximation
CJT 765: Structural Equation Modeling
Introduction to Instrumentation Engineering
Collaborative Filtering Matrix Factorization Approach
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Simple Linear Regression
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
Presentation transcript:

Semi-mechanistic modelling in Nonlinear Regression: a case study by Katarina Domijan 1, Murray Jorgensen 2 and Jeff Reid 3 1 AgResearch Ruakura 2 University of Waikato 3 Crop and Food Research

Introduction  Reid (2002) developed a crop response model  genetic algorithm  no measures of confidence for individual parameter estimates

General structure  non-linear model  semi-mechanistic  parts of the model that would ideally be mechanistic are replaced by empirically estimated functions  relatively complex  (26 parameters to be estimated)  generality of application  challenging to fit

General structure water stress, plant density, quantity of light etc yield under ideal nutrient/pH conditions (maximum yield) nutrient supply (N, P, K, Mg) observed yield These effects are assumed to act multiplicatively and independently of each other and the nutrient effects

General structure – maize data Crop hybrid

General structure - nutrients Model structure is the same for all nutrients Each nutrient supply is assumed to have: a minimum value (the crop yield is zero) and an optimal value (further increases cause no additional yield) Reid (2002) defines a scaled nutrient supply index which is: = 0 at the minimum nutrient value and = 1 at the optimal nutrient value In soil added as fertilizer nutrient supply efficiency factor 1 efficiency factor 2 Proportion of the optimum amount of nutrient supply

General structure - nutrients For each nutrient, the effect of scaled nutrient supply index (x) on yield is modelled using the family of curves: N opt where: γ =shape parameter

General structure - combining nutrients The combined scaled yield is given by: (or 0 if this is negative). Note:  are scaled yields corrected for unavailability of the respective nutrients  Nutrient stresses are assumed to affect yield independently of each other Soil pH  Treated as if it were an extra nutrient  Only stress due to low pH is modelled and not stress due to excessive pH if the effects of a particular stressor are known to be absent for a set of data, then

just 2 nutrients (N and K) Scaled yield General structure - combining nutrients

Data  model was tested for maize crops grown in the North Island between 1996 and 1999  data was collated from 3 different sources of measurements  experimental and commercial crops  12 sites  6 hybrids  84 observations

Genetic Algorithms  stochastic optimization tools that work on “Darwinian” models of population biology  don’t need requirement of differentiability!  relatively robust to local minima/maxima  don’t need initial values  have no indication of how well the algorithm has performed  convergence to a global optimum in a fixed number of generations?  slow to move from an arbitrary point in the neighbourhood of the global optimum to the optimum point itself  no measure of confidence for individual parameters

Our approaches:  simplifying the model:  1 nutrient (N)  9 parameters  simulated data  combining GA with derivative based methods:  common methods (Gauss-Newton, Levenberg- Marquardt)  AD MODEL BUILDER  obtain CI’s:  gradient information  likelihood methods

Correlation of Parameter Estimates: g Nm Np d b e1 e2 E.1 Nmin B 1 Nopt.. 1 delta 1 beta, 1 eta1. 1 eta2 1 E.n1, 1 E.n2, + Simulated data - simple model  investigate the structure of the correlation matrix  generated so it mimics the “real” data as much as possible  large n (300), small residual variance (0.01) N min and γ N are highly correlated! (blank) , * B Key:

Complete model n=50000, σ 2 = Key: (blank) , * B0.95-1

N min Levenberg-Marquardt algorithm  Maize data  use GA estimates as starting values  simple model:  multicollinearity  parameter Nmin tends to –ve  reparametrization

 complete model:  (again) biological restrictions (N min, K min =0)  problems with equations which are constant for ranges of values (eg scaled yields)  replace nondiffentiable functions (pH, water stress)  some stressors held constant (P, Mg) N opt constant Levenberg-Marquardt algorithm

 2 nutrients (N and K) + stress due to low pH + water and population stresses  12 parameters

Profile likelihood CI’s N opt estimate Wald CI Likelihood CI Approach outlined in Bates and Watts (1989) Assess validity of the linear approximation to the expectation surface

Profile likelihood CI’s Estimation surface seems to be nonlinear with respect to most of the parameters in the model Especially E N1 and pH c -> one sided CIs Better estimates of uncertainty than linear approx. results

AD Model builder  automatic differentiation  faster  observed information matrix (better se’s)  we run into the same problems as with L-M  requires model to be differentiable  good initial values

In the end... CI’s are too wide to be of ‘practical’ use e.g. for parameter N opt (optimum amt of N supply per tonne of maximum yield) : but in the ‘maize dataset’ N supply per tonne of maximum yield varies between 6 and 54 Problems of  nonidentifiability  correlated estimates  poor precision of estimation in certain directions These phenomena are not clearly distinguished in nonlinear setting L-M and ADMB estimate 95% LINEAR APPROX. CI’s 95% PROFILE LIKELIHOOD CI’s 95% CI’s (ADMB) (3.80, 33.35)(11.67, 31.45)(11.32, 25.83)

Recommendations  do more experimentation - collect more information about parameters  particularly ‘approximately nonidentifiable’ parameters  replace all nondifferentiable equations in the model with smooth versions  bootstrapping  global optimum?