DATA ANALYSIS ERRORS IN CHEMICAL ANALYSIS Normal phrases in describing results of an analysis “pretty sure” “very sure” “most likely” “improbable” Replaced.

Slides:



Advertisements
Similar presentations
DATA & STATISTICS 101 Presented by Stu Nagourney NJDEP, OQA.
Advertisements

“Students” t-test.
Chapter 7 Statistical Data Treatment and Evaluation
BPS - 5th Ed. Chapter 241 One-Way Analysis of Variance: Comparing Several Means.
CHAPTER 25: One-Way Analysis of Variance Comparing Several Means
Errors in Chemical Analyses: Assessing the Quality of Results
Chapter 7: Statistical Analysis Evaluating the Data.
Chapters 3 Uncertainty January 30, 2007 Lec_3.
Statistical Quality Control Lab 202 By S. O. Duffuaa Systems Engineering Department.
Quality Control Procedures put into place to monitor the performance of a laboratory test with regard to accuracy and precision.
PSY 1950 Confidence and Power December, Requisite Quote “The picturing of data allows us to be sensitive not only to the multiple hypotheses that.
How Science Works Glossary AS Level. Accuracy An accurate measurement is one which is close to the true value.
Inferences About Process Quality
Statistical Treatment of Data Significant Figures : number of digits know with certainty + the first in doubt. Rounding off: use the same number of significant.
Chapter 9: Introduction to the t statistic
Richard M. Jacobs, OSA, Ph.D.
ANALYTICAL CHEMISTRY CHEM 3811
Statistics Introduction 1.)All measurements contain random error  results always have some uncertainty 2.)Uncertainty are used to determine if two or.
Chapter 6 Random Error The Nature of Random Errors
Statistical Analysis Statistical Analysis
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 4.
Chapter 3 Scientific Measurement 3.1 Using and Expressing Measurements
Statistical Techniques I EXST7005 Conceptual Intro to ANOVA.
Principles of Analytical Chemistry (F13I11) Recommended textbook: “Fundamentals of Analytical Chemistry” Skoog, West and Holler, 7th Ed., 1996 (Saunders.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Statistical Inferences Based on Two Samples Chapter 9.
Estimation of Statistical Parameters
Chapter 2 Measurement & Problem Solving. Uncertainty There is a certain amount of doubt in every measurement – It is important to know the uncertainty.
Chem. 31 – 9/23 Lecture Guest Lecture Dr. Roy Dixon.
Statistics and Quantitative Analysis Chemistry 321, Summer 2014.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Version 2012 Updated on Copyright © All rights reserved Dong-Sun Lee, Prof., Ph.D. Chemistry, Seoul Women’s University Chapter 5 Errors in Chemical.
University of Ottawa - Bio 4118 – Applied Biostatistics © Antoine Morin and Scott Findlay 08/10/ :23 PM 1 Some basic statistical concepts, statistics.
Chapter 5 Errors In Chemical Analyses Mean, arithmetic mean, and average (x) are synonyms for the quantity obtained by dividing the sum of replicate measurements.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Overview CONFIDENCE INTERVALS STUDENT’S T / T STATISTICS
Lecture 4 Basic Statistics Dr. A.K.M. Shafiqul Islam School of Bioprocess Engineering University Malaysia Perlis
I Introductory Material A. Mathematical Concepts Scientific Notation and Significant Figures.
Introduction to Analytical Chemistry
Uncertainty & Error “Science is what we have learned about how to keep from fooling ourselves.” ― Richard P. FeynmanRichard P. Feynman.
Uncertainty and Error in Measurement (IB text - Ch 11) (If reviewing this slide in the senior year, there is also uncertainty information in the AP text.
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
Introduction to Inference: Confidence Intervals and Hypothesis Testing Presentation 8 First Part.
RESEARCH & DATA ANALYSIS
Analysis of Experimental Data; Introduction
Quality Control: Analysis Of Data Pawan Angra MS Division of Laboratory Systems Public Health Practice Program Office Centers for Disease Control and.
CfE Advanced Higher Physics
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Measurements and Their Analysis. Introduction Note that in this chapter, we are talking about multiple measurements of the same quantity Numerical analysis.
ERT 207 Analytical Chemistry ERT 207 ANALYTICAL CHEMISTRY Dr. Saleha Shamsudin.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
Chapter 13 Understanding research results: statistical inference.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Chapter 7: The Distribution of Sample Means
Hypothesis Tests. An Hypothesis is a guess about a situation that can be tested, and the test outcome can be either true or false. –The Null Hypothesis.
Chapter 6: Random Errors in Chemical Analysis. 6A The nature of random errors Random, or indeterminate, errors can never be totally eliminated and are.
Chapter 5: Errors in Chemical Analysis. Errors are caused by faulty calibrations or standardizations or by random variations and uncertainties in results.
Home Reading Skoog et al. Fundamental of Analytical Chemistry. Chapters 5 and 6.
3.1 Using and Expressing Measurements > 1 Copyright © Pearson Education, Inc., or its affiliates. All Rights Reserved. Chapter 3 Scientific Measurement.
Chapter 9 Introduction to the t Statistic
Topic 11 Measurement and data processing
Virtual University of Pakistan
ESTIMATION.
Two-Sample Hypothesis Testing
Lecture Slides Elementary Statistics Twelfth Edition
Redox QUANTITATIVE ANALYTICAL CHEMISTRY Electrochemistry
Satish Pradhan Dnyanasadhana College, Thane. Department of Chemistry S
Dnyanasadhana College, Thane. Department of Chemistry T. Y. B. Sc
What are their purposes? What kinds?
Lecture Slides Elementary Statistics Twelfth Edition
Chapter 5: Errors in Chemical Analysis
Presentation transcript:

DATA ANALYSIS ERRORS IN CHEMICAL ANALYSIS Normal phrases in describing results of an analysis “pretty sure” “very sure” “most likely” “improbable” Replaced by using mathematical statistical tests. Is there such a thing as “ERROR FREE ANALYSIS”? Impossible to eliminate errors. Can only be minimized. Can only be approximated to an acceptable precision. How reliable are our data TO OVERCOME ERRORS. Carry out replicate measurements. Analyse accurately known standards. Perform statistical tests on data. ?

Mean/Average Xi = individual values of x N = number of replicate measurements Median Data in the middle if the number is odd, arranged in ascending order. The average of two data in the middle if he number is even arranged in ascending order. Errors Absolute Error Absolute error = measured value – true value Absolute error, Relative error = Percent relative error = STATISTICS Range The different between the highest and lowest result.

Standard Deviation (SD), s Measure of the precision of a population of data. Small sample size N-1 = Degree of freedom Population (infinity) N = Number of replicate More smaller SD means more precise the analysis

Varian, V The square of standard deviation. For sample, V = s 2 For population, V =  2 Relative Standard Deviation (RSD) And Covarian (CV)

PRECISION Relates to reproducibility or repeatability of a result. How similar are values obtained in exactly the same way? Useful for measuring deviation from the mean. ACCURACY Measurement of agreement between experimental mean and true value (which may not be known!).

DIFFERENCE BETWEEN ACCURACY AND PRECISION Good precision does not guarantee accuracy.

High Precision High accuracy High precision Low accuracy Good precision does not guarantee accuracy. Low precision Low accuracy Low precision High accuracy

TYPES OF ERROR IN EXPERIMENTAL DATA Gross Errors Serious but very seldom occur in analysis. Usually obvious - give outlier readings. Detectable by carrying out sufficient replicate measurements. Experiments must be repeated. e.g. Instrument faulty, contaminate reagent. Random (indeterminate) Error Data scattered approximately symmetrically about a mean value. Affects precision, can only be controlled. Dealt with statistically. e.g. physical and chemical variables. Systematic (determinate) Error Determinable and that presumably can be either avoided or corrected. Several possible sources. Readings all too high or too low. Causing bias in technique. Can either be +ve or -ve. Affects accuracy.

SOURCES OF SYSTEMATIC ERROR Instrument Error Need frequent calibration - both for apparatus such as volumetric flasks, burettes etc., but also for electronic devices such as spectrometers. Examples: Fluctuation in power supply Temperature changes Method Error Due to inadequacies in physical or chemical behaviour of reagents or reactions (e.g. slow or incomplete reactions). Difficult to detect and the most serious systematic error. Example: Small excess of reagent required causing an indicator to undergo colour change that signal the completion of a reaction. Personal Error Sources: Physical handicap, prejudice, not competence. Examples: Insensitivity to colour changes Tendency to estimate scale readings to improve precision Preconceived idea of “true” value.

SYSTEMATIC ERRORS Systematic errors can be: Constant (e.g. error in burette reading - less important for larger values of reading). Proportional (e.g. presence of given proportion of interfering impurity in sample; equally significant for all values of measurement). Minimise Errors: Minimise instrument errors by careful recalibration and good maintenance of equipment. Minimise personal errors by care and self-discipline. Method errors - most difficult. “True” value may not be known. Three approaches to minimise: Analysis of certified standards (SRM) Use 2 or more independent methods Analysis of blanks

SIGNIFICANT FIGURES Minimum number of digits written in scientific notation without a loss in accuracy. Zero is significant only when, It occurs in the middle of a number significant figures significant figures It is the last number to the right of the decimal point significant figures 6.00  significant figures significant figures Addition-Subtraction Use the same number of decimal places as the number with the fewest decimal places = = 13.6 (1 dp) (3 dp) (2 dp) (1 dp) Multiplication - Division Use the same number of digits as the number with the fewest number of digits.

USE OF STATISTICS IN DATA EVALUATION Defining the interval of values around a set mean () within which the population mean (  ) can be expected with a given probability. The intervals are called confidence limits. Determining the number of replicates required to assure (at a desired probability) that an experimental mean () falls within a predicted interval of values around the population mean (  ). Estimating the probability that the experimental mean () and true value (  ) are different or two experimental mean are different (t test). Estimating the probability that data from two experiments are different in precision (F test). Deciding when to accept/reject outliers among replicates (Q test). Treating calibration data. Quality control.

CONFIDENCE LIMITS AND CONFIDENCE INTERVAL Confidence Limits Interval around the mean that probably contains . Extream value of, a < < b Confidence Interval The magnitude of the confidence limits.  Confidence Limits Confidence Level Fixes the level of probability that the mean is within the confidence limits. 99.7%, 99%, 95%, 90%, 80%, 68%, 50%

CONFIDENCE LIMITS (CL) SINCE THE EXACT VALUE OF POPULATION MEAN,  CANNOT BE DETERMINED, ONE MUST USE STATISTICAL THEORY TO SET LIMITS AROUND THE MEASURED MEAN,, THAT PROBABLY CONTAIN . CL ONLY HAVE MEANING WITH THE MEASURED STANDARD DEVIATION, S, IS A GOOD APPROXIMATION OF THE POPULATION STANDARD DEVIATION, , AND THERE IS NO BIAS IN THE MEASUREMENT. CL WHEN  IS KNOWN (POPULATION), N = Number of measurements

Values for z at various confidence levels are found in Table 1. Confidence Level, %z VALUES FOR Z

Examples: or,

CL For Small Data Set (N 20),  not known, t s Values of t depend on degree of freedom, (N - 1) and confidence level (from Table t). t also known as ‘student’s t’ and will be used in hypothesis test.

VALUES OF t AT VARIOUS CONFIDENCE LEVEL Degree of Freedo m Confidence Level 90 %95 %99 % infinity

Example: Data for the analysis of calcium in rock: 14.35%, 14.41%, 14.40%, 14.32% and 14.37%. Calculate the confidence interval at 95% confidence level. Min, = Standard deviation, s = From the table, At 95 % confidence level, N - 1 = 4, t = Confidence interval,

At different confidence level, Confidence LevelConfidence Interval 90%  = 14.37%  %  = 14.37%  %  = 14.37%  0.08 Summary: If the confidence level increased, the confidence interval is also increased. The probabilities of  appear in the interval increased

Example: Data for the analysis of calcium in rock: 14.35%, 14.41%, 14.40%, 14.32%, 14.45%, 14.50%, 14.25% and 14.37%. Calculate the confidence interval.

OTHER USAGE OF CONFIDENCE INTERVAL To determine number of replicates needed for the mean to be within the confidence interval. To determine systematic error. TO DETERMINE NUMBER OF REPLICATES  If  is known (s   ), If  is unknown

Example: Calculate the number of replicates needed to reduce the confidence interval to 1.5  g/mL at 95% confidence level. Given, s = 2.4  g/mL. At 95% confidence level, t = 1.96,

TO DETERMINE SYSTEMATIC ERROR Example A standard solution gave an absorption reading of at a particular wavelength. Ten measurements were done on a sample and the mean gave a value of 0.461, with standard deviation of Show whether systematic error exists in the measurements at 95% confidence level. Answer At 95% confidence level, N – 1 = 9, t = 2.26, The calculation gives confidence limit of, <  < Does the true mean belong to the interval? Does systematic error present?

Observations Hypothesis Model Valid?Reject Basis for further experiments YES NO TESTING A HYPOTHESIS

Hypothesis  Data  Theory Normally, the measured data is not always the same and seldom agree. We use statistics to test the disagreement. NULL HYPOTHESIS Null Hypothesis, H o – the values of two measured quantities do not differ (significantly) UNLESS we can prove it that the two values are significantly different. “Innocent until proven guilty” In Null Hypothesis, the calculated value of a parameter from the equation is compared to the parameter value from the table. If the calculated value is smaller than the table value, the hypothesis is accepted and vice-versa.

Null Hypothesis can be used to: Compare  and Compare and from two sets of data Compare s 1 and s 2 or  1 and  2 Compare s and 

t TEST Comparison between experimental mean and true mean (  and ) – the presence of systematic error Steps: 1) If  is not known, If  is known,

ii)Calculate t or z (t calc ) from the data. iii)Compare t calc and t table iv)If t calc > t table Reject Null Hypothesis (H o ) i.e. The different is due to systematic error. v)If t calc < t table Accept Null Hypothesis (H o ) i.e. ( ) The different is due to random error. t test

Example: The sulphur content of a sample of kerosene was found to be 0.123%. A new method was used on the same sample and the following data is obtained: %S : 0.112; 0.118; 0.113; Show whether systematic error is present in the new method. Null Hypothesis, Ho :  = = 0.116%,  = 0.123% s = Since t calc > t table, Ho is rejected and the two means are significantly different and thus systematic error is present. t table = 3.18 (95 %, N-1 = 3)

Other Solution: Since, ) is significant and there is systematic error in the measurement. (i.e > ), H o is rejected, the difference ( ) is significant and there is systematic error in the measurement.

2.Comparing two mean values and Normally used to determine whether the two samples are identical or not. The difference in the mean of two sets of the same analysis will provide information on the similarity of the sample or the existence of random error. Data:, and s 1, s 2 Ho : = We want to test whether - = 0

Assume, Calculate the value of t; Compare t calc with ttable; if t calc < t table, H o is accepted. The pooled standard deviation, s p is calculated using: where, N 1, N 2 are numbers of data in sets 1 and 2 N s is the number of data sets

Example: The source of wine (the vineyard) is identified by determining the alcohol content of the different barrels. Barrel 1: 6 determinations, = 12.61% Barrel 2: 4 determinations, = 12.53% sp = 0.07% Prove that both wine sources are different??? H o : = From the t table at 95% confidence, with degrees of freedom – 2 = 8, t table is Since t calc is smaller than t table, Ho is accepted and the source of the wine statistically is the same.

COMPARING THE PRECISION OF TWO MEASUREMENTS (THE F-TEST) Is Method A more precise than Method B? Is there any significant difference between both methods? With the degree of freedom = N – 1 H o : the precision is identical; s 1 = s 2 Then, if F calc < F table,H o is accepted. Since the values of F (from table) are always greater than 1, the smaller variance (the more precise) always become the denominator. V 1 > V 2, so

Example: The determination of CO in a mixture of gases using the standard procedure gave an s value of 0.21 ppm. The method was modified twice giving s 1 of 0.15 (12 degrees of freedom) and s 2 of 0.12 (12 degrees of freedom) Are the two modified methods more precise than the standard? H o : s 1 = s std. and H o : s 2 = s std. In the standard method, s  and the degrees of freedom becomes infinity. Refer the F table: Numerator = , and denominator = 12; giving the critical F value of 2.30 Since F 1 < F table, Ho is accepted, While F 2 >F table ; so Ho is rejected.

H o : s 1 = s std. and H o : s 2 = s std. CONCLUSIONS

THE DIXON TEST OR THE Q TEST A way of detecting outlier, a data, which is statistically, does not belong to the set. Example: Data: 10.05, 10.10, 10.15, 10.05, 10.45, By inspection, seem to be out of the data normal range. It is easier to see it when the numbers are arranged in a decreasing or increasing order , 10.05, 10.10, 10.10, 10.15, Should this data (10.45) be eliminated, the mean will change from the original value!

where, x q is the questionable data x n is its nearest neighbour w is the difference between the highest and the lowest value (range). The Q expt or Q calc will be compared with the Q critical or Q table, and the null hypothesis is checked. = 0.75 Q critical (95%, n = 6) = Q expt > Q critical Data (10.45) can be rejected.

VALUE OF Q Number of Observations Confidence Level 90 %95 %99 %

Example: An analysis on calcite gave the following percentage of CaO: 55.45, 56.04, 56.23, 56.00, Arrange the data in order: 55.45, 56.00, 56.04, 56.08, Suspected data: OR Q table from 5 determinations, 95% = Since Q calc <Q table. Data cannot be rejected. Since the Q calc is = Q table, the data cannot be rejected. Number of Observat ions Confidence Level 90 %95 %99 %

Example: An analysis on calcite gave the following percentage of CaO: 55.45, 56.04, 56.80, 56.23, 56.00,55.30, 55.08, and Any data should be rejected at 90% confidence level? Number of Observatio ns Confidence Level 90 %95 %99 %