"I was still a couple of miles above the clouds when it broke, and with such violence I fell to the ground that I found myself stunned, and in a hole nine.

Slides:



Advertisements
Similar presentations
SADC Course in Statistics Confidence intervals using CAST (Session 07)
Advertisements

Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: a Monte Carlo experiment Original citation: Dougherty, C. (2012) EC220.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Sampling Distributions (§ )
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Copyright © 2006 Pearson Addison-Wesley. All rights reserved. Lecture 4: Mathematical Tools for Econometrics Statistical Appendix (Chapter 3.1–3.2)
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
1 Sociology 601, Class 4: September 10, 2009 Chapter 4: Distributions Probability distributions (4.1) The normal probability distribution (4.2) Sampling.
SAMPLING DISTRIBUTIONS. SAMPLING VARIABILITY
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Chapter 7: Variation in repeated samples – Sampling distributions
Bootstrapping LING 572 Fei Xia 1/31/06.
Topic 3: Regression.
CHAPTER 8 Estimating with Confidence
1 We will now consider the distributional properties of OLS estimators in models with a lagged dependent variable. We will do so for the simplest such.
Random Variables and Probability Distributions
1 A MONTE CARLO EXPERIMENT In the previous slideshow, we saw that the error term is responsible for the variations of b 2 around its fixed component 
Bootstrapping applied to t-tests
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
A P STATISTICS LESSON 9 – 1 ( DAY 1 ) SAMPLING DISTRIBUTIONS.
Chapter 8 Introduction to Hypothesis Testing
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
T-test Mechanics. Z-score If we know the population mean and standard deviation, for any value of X we can compute a z-score Z-score tells us how far.
CHAPTER 8 Estimating with Confidence
Chapter 8 Introduction to Inference Target Goal: I can calculate the confidence interval for a population Estimating with Confidence 8.1a h.w: pg 481:
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 8: Estimating with Confidence Section 8.1 Confidence Intervals: The.
Biostatistics IV An introduction to bootstrap. 2 Getting something from nothing? In Rudolph Erich Raspe's tale, Baron Munchausen had, in one of his many.
Montecarlo Simulation LAB NOV ECON Montecarlo Simulations Monte Carlo simulation is a method of analysis based on artificially recreating.
Bootstrapping (And other statistical trickery). Reminder Of What We Do In Statistics Null Hypothesis Statistical Test Logic – Assume that the “no effect”
PARAMETRIC STATISTICAL INFERENCE
Fyzika tokamaků1: Úvod, opakování1 Tokamak Physics Jan Mlynář 6. Transport: Theory and experiments Bootstrap current, revision of transport, dimensional.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
Stat 112: Notes 2 Today’s class: Section 3.3. –Full description of simple linear regression model. –Checking the assumptions of the simple linear regression.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Section 10.1 Confidence Intervals
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Robust Estimators.
Copyright © 2009 Pearson Education, Inc. Chapter 19 Confidence Intervals for Proportions.
Chapter 10 The t Test for Two Independent Samples.
Inen 460 Lecture 2. Estimation (ch. 6,7) and Hypothesis Testing (ch.8) Two Important Aspects of Statistical Inference Point Estimation – Estimate an unknown.
1 ASYMPTOTIC PROPERTIES OF ESTIMATORS: THE USE OF SIMULATION In practice we deal with finite samples, not infinite ones. So why should we be interested.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
1 Probability and Statistics Confidence Intervals.
Nonlinear regression Review of Linear Regression.
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Non-parametric Approaches The Bootstrap. Non-parametric? Non-parametric or distribution-free tests have more lax and/or different assumptions Properties:
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Bootstrapping James G. Anderson, Ph.D. Purdue University.
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 8: Estimating with Confidence Section 8.1 Confidence Intervals: The.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Estimating standard error using bootstrap
Unit 5 – Chapters 10 and 12 What happens if we don’t know the values of population parameters like and ? Can we estimate their values somehow?
Inference for the Mean of a Population
Inference for Two-Samples
Chapter 3 INTERVAL ESTIMATES
Confidence Intervals for Proportions
Simple Linear Regression - Introduction
When we free ourselves of desire,
Neuroinformatics 1.1: the bootstrap
Distribution of the Sample Proportion
Chapter 9: Hypothesis Tests Based on a Single Sample
Chapter 8: Estimating With Confidence
Presentation transcript:

"I was still a couple of miles above the clouds when it broke, and with such violence I fell to the ground that I found myself stunned, and in a hole nine fathoms under the grass, when I recovered, hardly knowing how to get out again. Looking down, I observed that I had on a pair of boots with exceptionally sturdy straps. Grasping them firmly, I pulled with all my might. Soon I had hoist myself to the top and stepped out on terra firma without further ado.“ - Baron Munchausen (in Singular Travels, Campaigns and Adventures of Baron Munchausen, by R. E. Raspe, 1786.)

Introduction to Bootstrapping RIK CHAKRABORTI AND GAVIN ROBERTS

Sailing in the clouds

Sailing in the clouds – idealized assumptions and asymptotics

The cloud breaks…!

And we fall, hard! Previous example shows knowing the sampling distribution is important for valid inference. But, this may be difficult if – 1.Assumptions about the distribution of errors are false. Errors may not be distributed normally, or even asymptotically normally. 2.Computing sampling characteristics of certain statistics for finite sample sizes can be very difficult. Typically this is circumvented by resorting to asymptotic algebra. For example, for testing non-linear hypothesis when we use the delta method, we rely on asymptotic justification.

Stuck in a rut, 9 fathoms deep?

So, what do we do? Basic problem – have no clue about small sample properties of sampling distribution of estimator/statistic of interest. SOLUTION – GO MONTE CARLO?

Monte Carlo, with a difference! Typically, we’ve run Monte-Carlo simulations in the context of simple regressions. STEPS: 1.Simulate sample data using process that mimics true DGP. 2.Compute statistic of interest for sample. 3.Repeat for mind bogglingly large number of times as long as it doesn’t boggle the computer’s mind - Generates sampling distribution of statistic of interest

Monte Carlo, with a difference!

Monte Carlo, with a difference (steps 2 and 3)

Here’s the difference

Procedure

Why does it work? d

But…

Advantages of the paired bootstrap Keeps error paired with original explanatory variable it was associated with. Implicitly employs true errors, true underlying parameters and preserves original functional form. Allows explanatory variables to vary across samples – assumption of non-stochastic regressors relaxed.

Common uses  Estimation of standard errors when these are hard to compute  Figuring out proper size of tests, i.e., type – I error rates.  Bias correction.

Caution – check sturdiness of straps before the haul! Bootstrapping performs better in estimating sampling distributions of “asymptotically pivotal” statistics – statistics whose sampling distribution does not depend on unknown population parameters. – sampling distribution of parameter estimates typically depend on population parameters. Instead, bootstrapped sampling distribution of the t-statistic converges faster.

Further references for prospective bootstrappers 1.Kennedy – Chapter 4, section 6, if you want to understand the bootstrap 2.Cameron and Trivedi – Chapter 11, if you want to do the bootstrap 3.MacKinnon (2006) – Uses and abuses to be wary of. 4.And most importantly, Watch “The adventures of Baron Munchausen” the awesome Terry Gilliam movie.