Hypothesis testing and confidence intervals by resampling by J. Kárász.

Slides:



Advertisements
Similar presentations
Review bootstrap and permutation
Advertisements

Estimation of Means and Proportions
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Sampling: Final and Initial Sample Size Determination
Sampling Distributions (§ )
Ka-fu Wong © 2003 Chap 8- 1 Dr. Ka-fu Wong ECON1003 Analysis of Economic Data.
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Bootstrapping in regular graphs
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
Overview of STAT 270 Ch 1-9 of Devore + Various Applications.
Bootstrapping LING 572 Fei Xia 1/31/06.
Introduction to Probability and Statistics Chapter 7 Sampling Distributions.
1 Inference About a Population Variance Sometimes we are interested in making inference about the variability of processes. Examples: –Investors use variance.
STAT 572: Bootstrap Project Group Members: Cindy Bothwell Erik Barry Erhardt Nina Greenberg Casey Richardson Zachary Taylor.
Bootstrap spatobotp ttaoospbr Hesterberger & Moore, chapter 16 1.
1 Summarizing Performance Data Confidence Intervals Important Easy to Difficult Warning: some mathematical content.
QBM117 - Business Statistics Estimating the population mean , when the population variance  2, is unknown.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Bootstrapping (And other statistical trickery). Reminder Of What We Do In Statistics Null Hypothesis Statistical Test Logic – Assume that the “no effect”
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
CHAPTER 11 DAY 1. Assumptions for Inference About a Mean  Our data are a simple random sample (SRS) of size n from the population.  Observations from.
Chapter 7: Sample Variability Empirical Distribution of Sample Means.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
8 Sampling Distribution of the Mean Chapter8 p Sampling Distributions Population mean and standard deviation,  and   unknown Maximal Likelihood.
Resampling techniques
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Confidence Intervals (Dr. Monticino). Assignment Sheet  Read Chapter 21  Assignment # 14 (Due Monday May 2 nd )  Chapter 21 Exercise Set A: 1,2,3,7.
"Classical" Inference. Two simple inference scenarios Question 1: Are we in world A or world B?
1 Summarizing Performance Data Confidence Intervals Important Easy to Difficult Warning: some mathematical content.
Review Normal Distributions –Draw a picture. –Convert to standard normal (if necessary) –Use the binomial tables to look up the value. –In the case of.
Introduction to Inference Sampling Distributions.
8.1 Estimating µ with large samples Large sample: n > 30 Error of estimate – the magnitude of the difference between the point estimate and the true parameter.
Incorporating Uncertainties into Economic Forecasts: an Application to Forecasting Economic Activity in Croatia Dario Rukelj Ministry of Finance of the.
Chapter 9 Inferences Based on Two Samples: Confidence Intervals and Tests of Hypothesis.
Statistical Data Analysis 2011/2012 M. de Gunst Lecture 4.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Quantifying Uncertainty
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Statistics and probability Dr. Khaled Ismael Almghari Phone No:
Class Six Turn In: Chapter 15: 30, 32, 38, 44, 48, 50 Chapter 17: 28, 38, 44 For Class Seven: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 Read.
Estimating standard error using bootstrap
Inference: Conclusion with Confidence
Introduction For inference on the difference between the means of two populations, we need samples from both populations. The basic assumptions.
Statistical Estimation
Inference for the Mean of a Population
Standard Errors Beside reporting a value of a point estimate we should consider some indication of its precision. For this we usually quote standard error.
Chapter 6 Inferences Based on a Single Sample: Estimation with Confidence Intervals Slides for Optional Sections Section 7.5 Finite Population Correction.
ESTIMATION.
STAT 312 Chapter 7 - Statistical Intervals Based on a Single Sample
Inference: Conclusion with Confidence
Chapter Six Normal Curves and Sampling Probability Distributions
Parameter, Statistic and Random Samples
t distribution Suppose Z ~ N(0,1) independent of X ~ χ2(n). Then,
Test for Mean of a Non-Normal Population – small n
Simulation: Sensitivity, Bootstrap, and Power
Bootstrap - Example Suppose we have an estimator of a parameter and we want to express its accuracy by its standard error but its sampling distribution.
QQ Plot Quantile to Quantile Plot Quantile: QQ Plot:
Ch13 Empirical Methods.
Warmup To check the accuracy of a scale, a weight is weighed repeatedly. The scale readings are normally distributed with a standard deviation of
What are their purposes? What kinds?
Sampling Distributions (§ )
Fractional-Random-Weight Bootstrap
Bootstrapping and Bootstrapping Regression Models
Introductory Statistics
How Confident Are You?.
STATISTICS INFORMED DECISIONS USING DATA
Presentation transcript:

Hypothesis testing and confidence intervals by resampling by J. Kárász

Contents The bootstrap method The bootstrap analysis of the Kolmogorov- Smirnov test The bootstrap analysis of the GEV parameter investing The bootstrap analysis of the movig window method

Testing of homogenity Kolmogorov-Smirnov test Investigating the GEV parameter of dependence on time

Bootstrap method conditions of use is a random sample from the unknown probability distribution function (F) with finite variance. is the unknown parameter, the function of F. is the non-parametric estimate of the parameter, the function of the random sample.

Bootstrap method bootstrap estimate of the standard error is the standard error of the estimate. Then the bootstrap estimate is Unfortunatly in most cases it’s impossible to express it as a simple function of or the random sample, so we have to use numeric approximation.

Bootstrap method bootstrap sample and bootstrap replicate To approximate the empirical distribution function, the bootstrap algorithm takes random samples from the empirical distribution function: where This is the bootstrap sample. It is nothing else but a random sample from with replacement. By evaluating the statistic of interest we get a bootstrap replicate:

Bootstrap method approximation with Monte Carlo method 1.Independently draw a large number of bootstrap samples: 2.Evaluate the statistic of interest, so we get B bootstrap replicates: 3.Calculate the sample mean and sample standard deviation of the replicates:

Bootstrap method confidence intervals The histogram of the bootstrap replicates is an empirical density function for, so the  and  histogram percentiles are suitable limit estimates for the  percent confidence interval.

Bootstrap method hypotesis testing H 0 : are iid random variables. H 1 : are not iid random variables. The bootstrap samples are drawn from the same distribution ( ) independently, so if H 0 holds, then and the replicates are from quite similar distribution, because F and are similar for large sample size. If is out of the empirical confidence interval,we accept H 1, in other case, we accept H 0.

Kolmogorov-Smirnov test H 0 : are iid random variables. H 1 : are not iid random variables. Our suppose was that the annual maximum water levels are independent, so if we refuse H 0 we have to accept that are not from the same distribution. This can even mean trend.

Results 1 means H 0 was refused. t=1,2,3 : parameter for cutpoint in K-S test.

Results - example Annual maximum water level at Szolnok, t=1

Results - example Annual maximum water level at Szolnok, t=2

Examinating the GEV parameter of shape H 0 : are iid random variables. H 1 : are not iid random variables. Our suppose is that the annual maximum water levels are independent, so if we refuse H 0 we have to accept that are not from the same distribution. This can even mean trend.

Results 1 means H 0 was refused. No dataset was found to refuse H 0 both in K-S test and GEV parameter testing.

Results – data table

Results - example

Further questions – two-peeked bootstrap empirical distributions

Permutation testing similar to bootstrap method Same as the bootstrap algorithm except that permutation sample is drawn without replacement. The hypotesis testing is similar too, we examine the estimate and the empirical confidence interval.

Moving window method forecast analysis by permutation method Our aims were: Simulating the original dataset by permutation. Supervise the quality of the forecast.

Results -example