Information Criterion for Model Selection Romain Hugues.

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Ordinary Least-Squares
Structural Equation Modeling Using Mplus Chongming Yang Research Support Center FHSS College.
Statistical Techniques I EXST7005 Simple Linear Regression.
CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Model specification (identification) We already know about the sample autocorrelation function (SAC): Properties: Not unbiased (since a ratio between two.
Linear regression models
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
Aspects of Conditional Simulation and estimation of hydraulic conductivity in coastal aquifers" Luit Jan Slooten.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.
The Simple Linear Regression Model: Specification and Estimation
Model Fitting Jean-Yves Le Boudec 0. Contents 1 Virus Infection Data We would like to capture the growth of infected hosts (explanatory model) An exponential.
458 Model Uncertainty and Model Selection Fish 458, Lecture 13.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Chapter 4 Multiple Regression.
Multivariate Data Analysis Chapter 11 - Structural Equation Modeling.
Optimal Bandwidth Selection for MLS Surfaces
NOTES ON MULTIPLE REGRESSION USING MATRICES  Multiple Regression Tony E. Smith ESE 502: Spatial Data Analysis  Matrix Formulation of Regression  Applications.
SAMPLING DISTRIBUTIONS. SAMPLING VARIABILITY
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
This Week Continue with linear regression Begin multiple regression –Le 8.2 –C & S 9:A-E Handout: Class examples and assignment 3.
Logistic Regression with “Grouped” Data Lobster Survival by Size in a Tethering Experiment Source: E.B. Wilkinson, J.H. Grabowski, G.D. Sherwood, P.O.
Parallel implementation of RAndom SAmple Consensus (RANSAC) Adarsh Kowdle.
Stat13-lecture 25 regression (continued, SE, t and chi-square) Simple linear regression model: Y=  0 +  1 X +  Assumption :  is normal with mean 0.
Confidence Intervals for Means. point estimate – using a single value (or point) to approximate a population parameter. –the sample mean is the best point.
Chap 6-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 6 Introduction to Sampling.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 6-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
1 Chapter 3 Multiple Linear Regression Multiple Regression Models Suppose that the yield in pounds of conversion in a chemical process depends.
VI. Evaluate Model Fit Basic questions that modelers must address are: How well does the model fit the data? Do changes to a model, such as reparameterization,
H1H1 H1H1 HoHo Z = 0 Two Tailed test. Z score where 2.5% of the distribution lies in the tail: Z = Critical value for a two tailed test.
Chapter 14 Inference for Regression AP Statistics 14.1 – Inference about the Model 14.2 – Predictions and Conditions.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Ordinary Least Squares Regression.
Parameter estimation. 2D homography Given a set of (x i,x i ’), compute H (x i ’=Hx i ) 3D to 2D camera projection Given a set of (X i,x i ), compute.
Population Pharmacokinetic Characteristics of Levosulpiride and Terbinafine in Healthy Male Korean Volunteers Yong-Bok Lee College of Pharmacy and Institute.
Measurement Models: Identification and Estimation James G. Anderson, Ph.D. Purdue University.
Quadratic Classifiers (QC) J.-S. Roger Jang ( 張智星 ) CS Dept., National Taiwan Univ Scientific Computing.
A Flexible New Technique for Camera Calibration Zhengyou Zhang Sung Huh CSPS 643 Individual Presentation 1 February 25,
Properties of Estimators Statistics: 1.Sufficiency 2.Un-biased 3.Resistance 4.Efficiency Parameters:Describe the population Describe samples. But we use.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 11: Models Marshall University Genomics Core Facility.
How Good is a Model? How much information does AIC give us? –Model 1: 3124 –Model 2: 2932 –Model 3: 2968 –Model 4: 3204 –Model 5: 5436.
ES 07 These slides can be found at optimized for Windows)
Information criteria What function fits best? The more free parameters a model has the higher will be R 2. The more parsimonious a model is the lesser.
CWR 6536 Stochastic Subsurface Hydrology
Parameter estimation class 5 Multiple View Geometry CPSC 689 Slides modified from Marc Pollefeys’ Comp
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Computacion Inteligente Least-Square Methods for System Identification.
Relative and Absolute Fit Evaluation in Cognitive Diagnostic Modeling 23/9/2013.
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Chapter 9 Day 2. Warm-up  If students picked numbers completely at random from the numbers 1 to 20, the proportion of times that the number 7 would be.
Model Comparison. Assessing alternative models We don’t ask “Is the model right or wrong?” We ask “Do the data support a model more than a competing model?”
Evolutionary Computation: Advanced Algorithms and Operators
(5) Notes on the Least Squares Estimate
Chapter 4. Inference about Process Quality
The Simple Linear Regression Model: Specification and Estimation
LINEAR AND NONLINEAR MODELS
How Good is a Model? How much information does AIC give us?
Parameter estimation class 5
10701 / Machine Learning Today: - Cross validation,
Cross-validation for the selection of statistical models
Chapter 14 Inference for Regression
Introduction to Sensor Interpretation
Logistic Regression with “Grouped” Data
Introduction to Sensor Interpretation
Number of treatments ____________________________________________
Presentation transcript:

Information Criterion for Model Selection Romain Hugues

Problem description DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Important parameters We have N m-vector data a sampled from a m’-dim manifold A. We want to estimate a d-dim manifold S. S is parameterized by a n-vector u constrained to be in a n’-dim manifold U DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Model can be described by: d: dimension r ( = m’-d) : codimension n’ :degrees of freedom DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Minimization and expected residual Max. Lik. Solution of problem by minimizing J: New Notation for residual with respect to model: Residuals for future data a* : Expected residual of Model S: DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Mahalanobis projection of Data: DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Optimally fitted Manifold: DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Evaluation of expected residual: WE NEED TO ESTIMATE I(S) DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Geometric Information Criterion AIC(S) is an unbiased estimator of I(S): Extracting noise level ε from covariance: Normalized residual : Normalized AIC : DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Model Selection S 1 ”better” than S 2 if AIC 0 (S 1 ) <AIC 0 (S 2 ) If model S1 is CORRECT DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Model Comparison S 1 ”better” than S 2 if AIC 0 (S 1 ) <AIC 0 (S 2 ) DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

What should be done in practice? 1.Collect Data. 2.Estimate Manifolds and true positions for each model. 3.Compute Residuals for each model. 4.If a model is always “correct”, estimate noise level from residuals of this model 5.Compare two models: DESCRIPTIONTHEORETICAL BACKGROUNDIN PRACTICE

Situations when this is useful ? ??