Bootstrap for Goodness of Fit

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

STATISTICS HYPOTHESES TEST (III) Nonparametric Goodness-of-fit (GOF) tests Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering.
Regression Eric Feigelson Lecture and R tutorial Arcetri Observatory April 2014.
Hypothesis testing and confidence intervals by resampling by J. Kárász.
X-ray Astrostatistics Bayesian Methods in Data Analysis Aneta Siemiginowska Vinay Kashyap and CHASC Jeremy Drake, Nov.2005.
Pattern Recognition and Machine Learning
Probability and Statistics Basic concepts II (from a physicist point of view) Benoit CLEMENT – Université J. Fourier / LPSC
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Nonparametric density estimation or Smoothing the data Eric Feigelson Arcetri Observatory, April 2014.
A gentle introduction to Gaussian distribution. Review Random variable Coin flip experiment X = 0X = 1 X: Random variable.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Parametric Inference.
Kernel Methods Part 2 Bing Han June 26, Local Likelihood Logistic Regression.
7. Nonparametric inference  Quantile function Q  Inference on F  Confidence bands for F  Goodness- of- fit tests 1.
Regression Eric Feigelson. Classical regression model ``The expectation (mean) of the dependent (response) variable Y for a given value of the independent.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Simulation – Stat::Fit
An Empirical Likelihood Ratio Based Goodness-of-Fit Test for Two-parameter Weibull Distributions Presented by: Ms. Ratchadaporn Meksena Student ID:
Lecture note for Stat 231: Pattern Recognition and Machine Learning 4. Maximum Likelihood Prof. A.L. Yuille Stat 231. Fall 2004.
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Confidence Interval & Unbiased Estimator Review and Foreword.
Overview G. Jogesh Babu. Overview of Astrostatistics A brief description of modern astronomy & astrophysics. Many statistical concepts have their roots.
ES 07 These slides can be found at optimized for Windows)
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Statistical Data Analysis 2011/2012 M. de Gunst Lecture 4.
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Overview G. Jogesh Babu. R Programming environment Introduction to R programming language R is an integrated suite of software facilities for data manipulation,
Estimating standard error using bootstrap
Stat 223 Introduction to the Theory of Statistics
Advanced Data Analytics
Concepts in Probability, Statistics and Stochastic Modeling
Statistical Estimation
Chapter 4 Basic Estimation Techniques
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
HST 583 fMRI DATA ANALYSIS AND ACQUISITION
Why Stochastic Hydrology ?
Stat 223 Introduction to the Theory of Statistics
Ch8: Nonparametric Methods
The Maximum Likelihood Method
Overview G. Jogesh Babu.
Introduction to Instrumentation Engineering
Bootstrap - Example Suppose we have an estimator of a parameter and we want to express its accuracy by its standard error but its sampling distribution.
Graduierten-Kolleg RWTH Aachen February 2014 Glen Cowan
Model Comparison.
Data analysis in HEP: a statistical toolkit
BOOTSTRAPPING: LEARNING FROM THE SAMPLE
Stochastic Hydrology Hydrological Frequency Analysis (I) Fundamentals of HFA Prof. Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering.
POINT ESTIMATOR OF PARAMETERS
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Goodness-of-Fit Tests Applications
Lecture 3 1 Probability Definition, Bayes’ theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests general.
B.Mascialino, A.Pfeiffer, M.G.Pia, A.Ribon, P.Viarengo
Chapter 11 Bradt Notes ASTR 402 Dr. H. Geller
Stat 223 Introduction to the Theory of Statistics
Computing and Statistical Data Analysis / Stat 7
Validating a Random Number Generator
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and 2 Now, we need procedures to calculate  and 2 , themselves.
Parametric Methods Berlin Chen, 2005 References:
Introduction to Statistics − Day 4
UNIT-4.
Comparison of data distributions: the power of Goodness-of-Fit Tests
Stochastic Hydrology Fundamentals of Hydrological Frequency Analysis
Introductory Statistics
Uncertainty Propagation
Professor Ke-Sheng Cheng
Presentation transcript:

Bootstrap for Goodness of Fit G. Jogesh Babu Center for Astrostatistics http://astrostatistics.psu.edu

Astrophysical Inference from astronomical data Fitting astronomical data Non-linear regression Density (shape) estimation Parametric modeling Parameter estimation of assumed model Model selection to evaluate different models Nested (in quasar spectrum, should one add a broad absorption line BAL component to a power law continuum) Non-nested (is the quasar emission process a mixture of blackbodies or a power law?) Goodness of fit

Chandra X-ray Observatory ACIS data COUP source # 410 in Orion Nebula with 468 photons Fitting to binned data using c2 (XSPEC package) Thermal model with absorption, AV~1 mag

Fitting to unbinned EDF Maximum likelihood (C-statistic) Thermal model with absorption

Empirical Distribution Function

Incorrect model family Power law model, absorption AV~1 mag Question : Can a power law model be excluded with 99% confidence?

K-S Confidence bands F=Fn +/- Dn(a)

Model fitting Find most parsimonious `best’ fit to answer: Is the underlying nature of an X-ray stellar spectrum a non-thermal power law or a thermal gas with absorption? Are the fluctuations in the cosmic microwave background best fit by Big Bang models with dark energy or with quintessence? Are there interesting correlations among the properties of objects in any given class (e.g. the Fundamental Plane of elliptical galaxies), and what are the optimal analytical expressions of such correlations?

Statistics Based on EDF Kolmogrov-Smirnov: supx |Fn(x) - F(x)|, supx (Fn(x) - F(x))+, supx (Fn(x) - F(x))- Cramer - van Mises: Anderson - Darling: All of these statistics are distribution free Nonparametric statistics. But they are no longer distribution free if the parameters are estimated or the data is multivariate.

KS Probabilities are invalid when the model parameters are estimated from the data. Some astronomers use them incorrectly. (Lillifors 1964)

Multivariate Case Warning: K-S does not work in multidimensions Example – Paul B. Simpson (1951) F(x,y) = ax2 y + (1 – a) y2 x, 0 < x, y < 1 (X1, Y1) data from F, F1 EDF of (X1, Y1) P(| F1(x,y) - F(x,y)| < 0.72, for all x, y) is > 0.065 if a = 0, (F(x,y) = y2 x) < 0.058 if a = 0.5, (F(x,y) = xy(x+y)/2) Numerical Recipe’s treatment of a 2-dim KS test is mathematically invalid.

Processes with estimated Parameters {F(.; q): q e Q} - a family of distributions X1, …, Xn sample from F Kolmogorov-Smirnov, Cramer-von Mises etc., when q is estimated from the data, are Continuous functionals of the empirical process Yn (x; qn) = (Fn (x) – F(x; qn))

In the Gaussian case, q = (m,s2) and

Bootstrap Gn is an estimator of F, based on X1, …, Xn X1*, …, Xn* i.i.d. from Gn qn*= qn(X1*, …, Xn*) F(.; q) is Gaussian with q = (m, s2) and , then Parametric bootstrap if Gn =F(.; qn) X1*, …, Xn* i.i.d. from F(.; qn) Nonparametric bootstrap if Gn =Fn (EDF)

Parametric Bootstrap X1*, …, Xn* sample generated from F(.; qn). In Gaussian case . Both supx |Fn (x) – F(x; qn)| and supx |Fn* (x) – F(x; qn*)| have the same limiting distribution (In the XSPEC packages, the parametric bootstrap is command FAKEIT, which makes Monte Carlo simulation of specified spectral model)

Nonparametric Bootstrap X1*, …, Xn* i.i.d. from Fn. A bias correction Bn(x) = Fn (x) – F(x; qn) is needed. supx |Fn (x) – F(x; qn)| and supx |Fn* (x) – F(x; qn*) - Bn (x) | have the same limiting distribution (XSPEC does not provide a nonparametric bootstrap capability)

Chi-Square type statistics – (Babu, 1984, Statistics with linear combinations of chi-squares as weak limit. Sankhya, Series A, 46, 85-93.) U-statistics – (Arcones and Giné, 1992, On the bootstrap of U and V statistics. Ann. of Statist., 20, 655–674.)

Confidence limits under misspecification of model family X1, …, Xn data from unknown H. H may or may not belong to the family {F(.; q): q e Q}. H is closest to F(.; q0), in Kullback - Leibler information h(x) log (h(x)/f(x; q)) dn(x) 0 h(x) |log (h(x)| dn(x) < h(x) log f(x; q0) dn(x) = maxq h(x) log f(x; q) dn(x)

supx |Fn* (x) – F(x; qn*) – (Fn (x) – F(x; qn)) | For any 0 < a < 1, P( supx |Fn (x) – F(x; qn) – (H(x) – F(x; q0)) | <Ca*) a Ca* is the a-th quantile of supx |Fn* (x) – F(x; qn*) – (Fn (x) – F(x; qn)) | This provide an estimate of the distance between the true distribution and the family of distributions under consideration.

References G. J. Babu and C. R. Rao (1993). Handbook of Statistics, Vol 9, Chapter 19. G. J. Babu and C. R. Rao (2003). Confidence limits to the distance of the true distribution from a misspecified family by bootstrap.   J. Statist. Plann. Inference 115, 471-478. G. J. Babu and C. R. Rao (2004). Goodness-of-fit tests when parameters are estimated.   Sankhya, Series A, 66 (2004) no. 1, 63-74.

The End