4.0 Confidence Limits cdf for the z-statistic

Slides:



Advertisements
Similar presentations
Chap 8-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 8 Estimation: Single Population Statistics for Business and Economics.
Advertisements

Math 144 Confidence Interval.
Chapter 11- Confidence Intervals for Univariate Data Math 22 Introductory Statistics.
BCOR 1020 Business Statistics Lecture 17 – March 18, 2008.
Point estimation, interval estimation
Chapter 8 Estimation: Single Population
Fall 2006 – Fundamentals of Business Statistics 1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter 7 Estimating Population Values.
Inferences About Process Quality
Chapter 7 Estimating Population Values
8-1 Introduction In the previous chapter we illustrated how a parameter can be estimated from sample data. However, it is important to understand how.
Chapter 6 Random Error The Nature of Random Errors
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Copyright © Cengage Learning. All rights reserved. 12 Simple Linear Regression and Correlation.
Confidence Interval Estimation
Error Analysis Accuracy Closeness to the true value Measurement Accuracy – determines the closeness of the measured value to the true value Instrument.
Estimation of Statistical Parameters
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Estimation Bias, Standard Error and Sampling Distribution Estimation Bias, Standard Error and Sampling Distribution Topic 9.
PARAMETRIC STATISTICAL INFERENCE
Statistics and Quantitative Analysis Chemistry 321, Summer 2014.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Introduction to Biostatistics, Harvard Extension School © Scott Evans, Ph.D.1 Descriptive Statistics, The Normal Distribution, and Standardization.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
General Statistics Ch En 475 Unit Operations. Quantifying variables (i.e. answering a question with a number) 1. Directly measure the variable. - referred.
Chapter 5 Parameter estimation. What is sample inference? Distinguish between managerial & financial accounting. Understand how managers can use accounting.
Section 10.1 Confidence Intervals
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
Inferential Statistics Part 1 Chapter 8 P
Probability = Relative Frequency. Typical Distribution for a Discrete Variable.
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
Sample Size Determination Text, Section 3-7, pg. 101 FAQ in designed experiments (what’s the number of replicates to run?) Answer depends on lots of things;
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
ME Mechanical and Thermal Systems Lab Fall 2011 Chapter 3: Assessing and Presenting Experimental Data Professor: Sam Kassegne, PhD, PE.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 7-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
CHAPTER 2.3 PROBABILITY DISTRIBUTIONS. 2.3 GAUSSIAN OR NORMAL ERROR DISTRIBUTION  The Gaussian distribution is an approximation to the binomial distribution.
Cell#1 Program Instructions. Don’t run. Used to load a Statistical Package Cell #3 Defines standard normal pdf and cdf functions Ignore spelling warning.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Business Statistics: A First Course 5 th Edition.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
6-1 Copyright © 2014, 2011, and 2008 Pearson Education, Inc.
10.1 Estimating with Confidence Chapter 10 Introduction to Inference.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Dr.Theingi Community Medicine
Copyright © Cengage Learning. All rights reserved. 9 Inferences Based on Two Samples.
CHAPTER 8 Estimating with Confidence
Physics 114: Lecture 13 Probability Tests & Linear Fitting
ESTIMATION.
Introduction to inference Estimating with confidence
Hypotheses and test procedures
Other confidence intervals
Point and interval estimations of parameters of the normally up-diffused sign. Concept of statistical evaluation.
Sampling Distributions and Estimation
Statistical Methods For Engineers
CONCEPTS OF ESTIMATION
Confidence intervals: The basics
Confidence Interval Estimation
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
4.1 Hypothesis Testing z-test for a single value
5.2 Least-Squares Fit to a Straight Line
Functions.
3.3 Experimental Standard Deviation
2.1 Properties of PDFs mode median expectation values moments mean
Confidence intervals: The basics
CHAPTER – 1.2 UNCERTAINTIES IN MEASUREMENTS.
2/5/ Estimating a Population Mean.
Objectives 6.1 Estimating with confidence Statistical confidence
Objectives 6.1 Estimating with confidence Statistical confidence
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

4.0 Confidence Limits cdf for the z-statistic probability and confidence intervals confidence interval of the average some practical comments about the number of replicates visualizing confidence intervals confidence intervals with s the t-parameter and its pdf t-based confidence intervals replicates revisited 4.0 : 1/13

Normal CDF The normal pdf is given by one particular form of Gauss's equation. This equation does not have an indefinite integral, thus the cdf cannot be written in a closed form. The value of F(m) has to be determined by numeric integration, with the results tabulated. To avoid needing a table for all values of m and s, a normalized variable, called the z-statistic, is created. F(z=-3) = 0.001 F(z=0) = 0.500 F(z=+3) = 0.999 4.0 : 2/13

Probability Intervals The cdf can tell us the range about m that contains a specified fraction of all measured values. For a normal density, this range is ordinarily chosen to be symmetric about m. Using the tabulated values, probability can be determined for the listed intervals. -k < z < +k -1.00 < z < +1.00 68% -1.96 < z < +1.96 95% -2.00 < z < +2.00 95.4% -3.00 < z < +3.00 99.7% The above inequalities can be rewritten in terms of x by using the definition of z. 4.0 : 3/13

Confidence Intervals Probability intervals tell us the range in which we expect to find a certain percentage of measured random values. Of more interest would be the interval about a single measurement that would contain the mean, m. This range is called the confidence interval or confidence limits. x - ks < m < x + ks The confidence level used in analytical chemistry is 95%, thus we would use k = 1.96, and say that the interval from x - 1.96s to x + 1.96s contains the true value 95% of the time. To simplify the calculations, 1.96 is almost always rounded up to 2. With this change, the confidence range is written m = x  2s. It has to be recognized that 5% of the time we write such a range it will not contain m. If it were desired to increase confidence to 99.9% the range would increase to x  3.3s. This range is considered too large by the analytical community, in view of the fact that confidence has only increased ~5%. 4.0 : 4/13

Confidence Interval of the Average The confidence interval of an average can be written in a similar manner by remembering that if x has the pdf n(m,s), then the average, , has the pdf n(m,s/N1/2). For 95% confidence we have the following interval. Both the equation and the graph imply that, for a given confidence, the range can be arbitrarily decreased by averaging more data. 4.0 : 5/13

Comments on Increasing N the number of replicate measurements can only be increased under very strict conditions obtaining replicates requires more time, allowing instrumental sensitivity to drift reagent concentrations can change over time, e.g. NaOH for some experiments replicates cost money, e.g. titrations some samples are not amenable to replication, e.g. taking blood from a newborn child since the interval only drops as the root of the replicates, it is often more effective to decrease s When performing an analysis for the first time we usually take three replicates. This is the minimum number of replicates allowing a "vote," that is, recognize outliers. when s is unknown, it is often cost effective to use 10 to 12 replicates to decrease the uncertainty in s 4.0 : 6/13

Visualization of Confidence Intervals The data below are the confidence intervals for the averages of three values. The random variable comes from a normal density, n(5,0.1). The averages then have a normal density, n(5,0.058). The 95% confidence interval is given by m = avg  0.116. The replicates with ranges not including m are identified. This method of graphing ranges is used most often in curve fitting.    4.0 : 7/13

Control Charts In a control chart the average is plotted along with the expected upper and lower ends of the confidence range. With 60 replicates you would expect on average three of the data points to be outside the range. If more or less points are outside the range, the analysis is out of control. This plot is used extensively in quality control when analyzing calibration standards. 4.0 : 8/13

1.96 Confidence Interval with s When s is substituted for s to establish the confidence limits, m is not included 95% of the time. This is shown in the following graph, where 9 miss the mean instead of ~0.0560 = ~3.          4.0 : 9/13

Why Doesn't the Interval with s Work? The numeric limits we have been using are actually values of the zstatistic. This can be seen by rearranging the limit expression. The average has the normal pdf, n(m,s/N1/2); the numerator has the pdf, n(0,s/N1/2); and, z has the pdf, n(0,1). What happens when s is replaced by s? To determine this, use the symbol, t, for the numeric limit and look at its functional form. The numerator of t has the normal pdf, n(0,s/N1/2). However, the denominator has the pdf of s, which was shown in Section 3.3 to depend upon N and have the functional form of a normal moment. 4.0 : 10/13

The PDF of t The pdf of t is derived in a manner similar to that for the standard deviation. The graph of f(t) is shown at the right for N = 3 (red), N = 6 (blue) and N = 101 (magenta). The pdf is symmetric about 0. As N  , the pdf approaches n(0,1). Numeric integration of f(t) is used to find the value of t that has a probability interval of F(t) - F(-t) = 0.95. 4.0 : 11/13

Confidence Intervals Using t The t-limits are given in the following table for a few values of the degrees of freedom, f = N-1. f 2 3 4 9 19 29 39  t(0.95) 4.30 3.18 2.78 2.26 2.09 2.05 2.02 1.96 The 95% confidence limits about the average of 3 replicates is given by . The data set previously shown with 1.96 is now shown with 4.30.   4.0 : 12/13

Value of N when Using t-Intervals When s is used to determine confidence intervals it is often more advantageous to make replicate measurements than when using s. This is because the value of t is decreasing at the same time that 1/N1/2 is decreasing. The graph below shows both z/N1/2 and t/N1/2. With t-intervals it is often cost effective to run ~10 replicates. Much beyond this value, the reduction in interval size is dictated by the slowly changing 1/N1/2 term. In particular note the factor of ~2 reduction in going from N = 3 to N = 5. 4.0 : 13/13