Designing powerful sampling programs Dr. Dustin Marshall.

Slides:



Advertisements
Similar presentations
Sampling Design, Spatial Allocation, and Proposed Analyses Don Stevens Department of Statistics Oregon State University.
Advertisements

Analysis by design Statistics is involved in the analysis of data generated from an experiment. It is essential to spend time and effort in advance to.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
+ Experiments How to Experiment Well: The RandomizedComparative Experiment The remedy for confounding is to perform a comparative experiment in which some.
Statistics in Science  Role of Statistics in Research.
Experimental design in environmental assessment  Environmental sampling and analysis (Quinn & Keough, 2003)
Slide 1-1 Copyright © 2004 Pearson Education, Inc. Stats Starts Here Statistics gets a bad rap, and Statistics courses are not necessarily chosen as fun.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
The current status of fisheries stock assessment Mark Maunder Inter-American Tropical Tuna Commission (IATTC) Center for the Advancement of Population.
Sampling and Randomness
Slide Slide 1 Copyright © 2007 Pearson Education, Inc Publishing as Pearson Addison-Wesley. Created by Tom Wegleitner, Centreville, Virginia Section 1-3.
Inferential Statistics
Chapter 4 Hypothesis Testing, Power, and Control: A Review of the Basics.
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 13 Experiments and Observational Studies.
Random Sampling, Point Estimation and Maximum Likelihood.
PARAMETRIC STATISTICAL INFERENCE
 Is there a comparison? ◦ Are the groups really comparable?  Are the differences being reported real? ◦ Are they worth reporting? ◦ How much confidence.
Slide 13-1 Copyright © 2004 Pearson Education, Inc.
Quantitative Analysis. Quantitative / Formal Methods objective measurement systems graphical methods statistical procedures.
LT 4.2 Designing Experiments Thanks to James Jaszczak, American Nicaraguan School.
Sampling and Sample Size Part 1 Cally Ardington. Course Overview 1.What is Evaluation? 2.Outcomes, Impact, and Indicators 3.Why Randomise? 4.How to Randomise?
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 4: Designing Studies Section 4.2 Experiments.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Copyright  2003 by Dr. Gallimore, Wright State University Department of Biomedical, Industrial Engineering & Human Factors Engineering Human Factors Research.
Sampling And Resampling Risk Analysis for Water Resources Planning and Management Institute for Water Resources May 2007.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
CHAPTER 9: Producing Data: Experiments. Chapter 9 Concepts 2  Observation vs. Experiment  Subjects, Factors, Treatments  How to Experiment Badly 
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
How to Avoid the Lies and Damned Lies: Pitfalls of Data Analysis Clay Helberg Special Topics in Marketing Research Dr. Charles Trappey Summarized by Kevin.
C82MST Statistical Methods 2 - Lecture 1 1 Overview of Course Lecturers Dr Peter Bibby Prof Eamonn Ferguson Course Part I - Anova and related methods (Semester.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
Lecture 6 Your data and models are never perfect… Making choices in research design and analysis that you can defend.
BIOSTAT - 1 Data: What types of data do you deal with? What do you think “statistics” means? Where do you obtain your data? What is a random variable in.
Selecting a Sample. outline Difference between sampling in quantitative & qualitative research.
Producing Data 1.
WELCOME TO BIOSTATISTICS! WELCOME TO BIOSTATISTICS! Course content.
AP Statistics Part IV – Inference: Conclusions with Confidence Chapter 10: Introduction to Inference 10.1Estimating with Confidence To make an inference.
STA248 week 121 Bootstrap Test for Pairs of Means of a Non-Normal Population – small samples Suppose X 1, …, X n are iid from some distribution independent.
Lecture Notes and Electronic Presentations, © 2013 Dr. Kelly Significance and Sample Size Refresher Harrison W. Kelly III, Ph.D. Lecture # 3.
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Chapter 1 – Ecological Data
CHAPTER 4 Designing Studies
Components of Experiments
Statistical Analysis Error Bars
Chapter 4: Designing Studies
CHAPTER 4 Designing Studies
Chapter 4: Designing Studies
Statistical Reasoning December 8, 2015 Chapter 6.2
Experimental Design Data Normal Distribution
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
CHAPTER 4 Designing Studies
Chapter 4: Designing Studies
Maintenance Sheet Due Wednesday
CHAPTER 4 Designing Studies
Maintenance Sheet Due Wednesday
VARIABILITY IN TRIALS Adapted fr M Gunther.
Principles of Experimental Design
Measuring the Wealth of Nations
Advanced Tools and Techniques of Program Evaluation
CHAPTER 4 Designing Studies
Maintenance Sheet Due Wednesday
10/28/ B Experimental Design.
Rest of lecture 4 (Chapter 5: pg ) Statistical Inferences
Presentation transcript:

Designing powerful sampling programs Dr. Dustin Marshall

Review Confounding Replication Randomisation

Causation or correlation?

Differences between treatments Does this address the hypothesis? Farm originForest origin Moved to new placeMoved to a new place Now at forestNow at farm

One treatment could react differently to movement that the other  confounding

Differences between treatments What is the unit of replication? Farm originForest origin Moved to new placeMoved to a new place Now at forestNow at farm Moved to new place but same environment

Summary of how to do a transplant experiment Must have transplant controls – manipulation that causes same disturbance but doesn’t change the environment Must replicate You should be able to draw any transplant experiment with controls

The real world Limited funds/time Can’t sample everything everywhere Could sample at multiple scales but how do you allocate resources?

Different sources of variation “Real” Variability Error 1. Imperfect accuracy 2. Imprecision

Accuracy vs precision

So, three sources of variation Minimise the artefactual sources of variation (minimise inaccuracy) Get an accurate estimate of ‘real variation’ Maximise precision but not at the expense of accuracy

The basics of a powerful experiment design Var among Var within = Ratio N Big  significant

The basics of a powerful experiment design Var among Var within = Ratio N

The basics of a powerful experiment design Var among Var within = Ratio N

Design 1 Measure length of fish to the nearest micron Catch fish by chasing them Take 400 samples High precision but not accurate due to bias

Design II Measure fish to the nearest micron Sample from the population randomly Take 10 samples High precision but inaccurate due to few samples of highly variable population

Design III Measure fish to the nearest metre Sample from the population randomly Take 100 samples Very low precision and therefore poor ability to distinguish variation

Design IV Catch fish randomly Measure to the nearest mm Measure 500 Low precision but high accuracy

The ideal, powerful design High precision –careful measurement –Fancy equipment High accuracy –Unbiased sampling –Replication –Precision High Power –Minimise unexplained variation (more about this later)

Scales of sampling Subsampling can increase precision Replication increases accuracy Is there ever a reason to sample at higher scales?

Higher levels of sampling Sampling at higher spatial or temporal scales increases our confidence in the generality of our findings Increases the size of your sampling universe

Continuum of sampling scales Scale of replication SubsampleHigh level unitReplicate PrecisionAccuracyGenerality

A trade-off Accuracy Precision Generality

Example: Does pollution reduce the quality of offspring that coral produce in the GBR? Polluted reefs vs unpolluted reefs Corals occur in patches Each coral can spawn 1,000,000 eggs

What/where to sample to sample? Site Reef Patch Coral Egg What is the unit of replication?

Where/what to sample Site Reef Patch Coral Egg Putting most effort here will maximise n But with no replication here, Var within will go up  bad Putting effort here will increase generality

You must replicate at the scales you want to make inferences about! Not just in relation to treatment effects but also how far you would like to extrapolate (within reason)

When should you increase your subsamples? When you are estimating only a portion of an entity and that estimate could be unreliable

Time – the fourth dimension Whenever you are measuring change, you must replicate over space and time Deciding how to allocate effort to each can be tricky Depends on what you’re interested in

So we’ve learnt how to design our sampling program to maximise accuracy but does that automatically translate into a powerful design? Not necessarily

The basics of a powerful experiment design Var among Var within = Ratio N Big  significant

Getting accurate estimates of these elements is helpful, as is maximising ‘n’ Var among Var within = Ratio N

Summary Precision, Accuracy, Generality Blocking – using random factors to partition variation and reduce unexplained variation Power analyses – used to give confidence that there is no type II error