Comparison of Acceptance Criteria The acceptance rates of simulated samples with and without various problems were examined to compare different acceptance.

Slides:



Advertisements
Similar presentations
4/4/2015Slide 1 SOLVING THE PROBLEM A one-sample t-test of a population mean requires that the variable be quantitative. A one-sample test of a population.
Advertisements

Statistical Techniques I EXST7005 Start here Measures of Dispersion.
Hypothesis testing Another judgment method of sampling data.
Chapter 9 Hypothesis Testing Understandable Statistics Ninth Edition
Chapter 10 Quality Control McGraw-Hill/Irwin
T-tests Computing a t-test  the t statistic  the t distribution Measures of Effect Size  Confidence Intervals  Cohen’s d.
Introduction to Hypothesis Testing
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. 10 Quality Control.
1/55 EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2008 Chapter 10 Hypothesis Testing.
Chapter 9 Audit Sampling: An Application to Substantive Tests of Account Balances McGraw-Hill/Irwin ©2008 The McGraw-Hill Companies, All Rights Reserved.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Basic Business Statistics.
BCOR 1020 Business Statistics Lecture 21 – April 8, 2008.
2.3. Measures of Dispersion (Variation):
IENG 486 Statistical Quality & Process Control
Statistical Methods in Computer Science Hypothesis Testing I: Treatment experiment designs Ido Dagan.
BA 427 – Assurance and Attestation Services
Slide 9-1 © The McGraw-Hill Companies, Inc., 2006 Audit Sampling.
Chapter 10 Hypothesis Testing
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
Fundamentals of Hypothesis Testing: One-Sample Tests
Section 9.1 Introduction to Statistical Tests 9.1 / 1 Hypothesis testing is used to make decisions concerning the value of a parameter.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap th Lesson Introduction to Hypothesis Testing.
Psy B07 Chapter 8Slide 1 POWER. Psy B07 Chapter 8Slide 2 Chapter 4 flashback  Type I error is the probability of rejecting the null hypothesis when it.
Jan 17,  Hypothesis, Null hypothesis Research question Null is the hypothesis of “no relationship”  Normal Distribution Bell curve Standard normal.
Chapter 3 Descriptive Measures
1 Introduction to Hypothesis Testing. 2 What is a Hypothesis? A hypothesis is a claim A hypothesis is a claim (assumption) about a population parameter:
+ The Practice of Statistics, 4 th edition – For AP* STARNES, YATES, MOORE Chapter 9: Sampling Distributions Section 9.2 Sample Proportions.
Chapter 13 Statistical Quality Control Method
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
L Berkley Davis Copyright 2009 MER301: Engineering Reliability Lecture 9 1 MER301:Engineering Reliability LECTURE 9: Chapter 4: Decision Making for a Single.
RDPStatistical Methods in Scientific Research - Lecture 41 Lecture 4 Sample size determination 4.1 Criteria for sample size determination 4.2 Finding the.
Chap 8-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 8 Introduction to Hypothesis.
Lecture 9 Chap 9-1 Chapter 2b Fundamentals of Hypothesis Testing: One-Sample Tests.
Selecting Input Probability Distribution. Simulation Machine Simulation can be considered as an Engine with input and output as follows: Simulation Engine.
The Sampling Distribution of
Type I and Type II Errors. An analogy may help us to understand two types of errors we can make with inference. Consider the judicial system in the US.
Chapter 8: Confidence Intervals based on a Single Sample
Chap 8-1 Fundamentals of Hypothesis Testing: One-Sample Tests.
Understanding Basic Statistics Fourth Edition By Brase and Brase Prepared by: Lynn Smith Gloucester County College Chapter Nine Hypothesis Testing.
Sampling Theory and Some Important Sampling Distributions.
Education 793 Class Notes Inference and Hypothesis Testing Using the Normal Distribution 8 October 2003.
9-1 Copyright © 2016 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
CHAPTER 2: Basic Summary Statistics
Type I and Type II Errors. For type I and type II errors, we must know the null and alternate hypotheses. H 0 : µ = 40 The mean of the population is 40.
1 Outline 1. Hypothesis Tests – Introduction 2. Technical vocabulary Null Hypothesis Alternative Hypothesis α (alpha) β (beta) 3. Hypothesis Tests – Format.
Power of a test. power The power of a test (against a specific alternative value) Is In practice, we carry out the test in hope of showing that the null.
Quality Control Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill.
Hypothesis Tests. An Hypothesis is a guess about a situation that can be tested, and the test outcome can be either true or false. –The Null Hypothesis.
Statistics for Business and Economics 8 th Edition Chapter 7 Estimation: Single Population Copyright © 2013 Pearson Education, Inc. Publishing as Prentice.
Chapter 9 Hypothesis Testing Understanding Basic Statistics Fifth Edition By Brase and Brase Prepared by Jon Booze.
Chapter Nine Hypothesis Testing.
Review of Power of a Test
Chapter 9 Audit Sampling: An Application to Substantive Tests of Account Balances McGraw-Hill/Irwin ©2008 The McGraw-Hill Companies, All Rights Reserved.
Chapter 9 Audit Sampling 1.
Understanding Results
Chapter 7: Sampling Distributions
Statistical Process Control
Chapter 9 Hypothesis Testing.
Daniela Stan Raicu School of CTI, DePaul University
Chapter Nine Part 1 (Sections 9.1 & 9.2) Hypothesis Testing
Section 2.5 notes Measures of Variation
When You See (This), You Think (That)
Another Example Consider a very popular computer game played by millions of people all over the world. The average score of the game is known to be
Chapter 7: Sampling Distributions
Daniela Stan Raicu School of CTI, DePaul University
1/10/ Sample Proportions.
Chapter 9 Hypothesis Testing: Single Population
Power and Error What is it?.
2.3. Measures of Dispersion (Variation):
Presentation transcript:

Comparison of Acceptance Criteria The acceptance rates of simulated samples with and without various problems were examined to compare different acceptance criteria E. Clarkson National Center for Advanced Materials Performance Presented to: Mil-HDBK-17 NCAMP December 2005, Santa Monica CA

Qualifying Sample Variables Sample size –n = 18 & n = 30 Mean –Set at 100 –Also tested at the 5 th and 95 th percentiles to see the effect of an unusually high or low qualifying sample mean. Coefficient of variation (std dev / mean) –2%, 6%, and 10% –Since the mean is set to 100, the Coefficient of Variation in % and the standard deviation have the same value.

Equivalence Sample Variables 50,000 samples generated for each set of conditions. Sample size –Tested at n = 2 to n = 10 Mean –Set at 100, at 95, and at 100 – 1σ. (σ = standard deviation) –Due to random generation of sample, the means will randomly vary around the preset value. Coefficient of variation (C.V.) –Set at 2%, 6%, and 10%, –Set at 1.5*C.V. of Qualifying sample. –Due to random generation of sample, the C.V. will randomly vary around the preset value.

Acceptance Criteria Agate Criteria: –Detailed in DOT/FAA/AR-03/19 “Material Qualification and equivalency for Polymer Matrix Composite Material Systems: Updated Procedure”, Final Report Sept 2003 with alpha (α) set at Cpk Criteria –A minimum Cpk of 1.33 is desired. The needed sample mean and minimum values are computed from the qualifying sample values based on that criteria. Criteria presented at Sept 1994 Mil-Hdbk-17 meeting. –A total of 8 different acceptance criteria used at that time by companies were tested to see how they performed compared to the recommended (AGATE) criteria.

Acceptance Criteria Criteria presented at Sept 1994 Mil-Hdbk-17 meeting 1.Sample Average > B-Basis, Sample Minimum > A-basis 2.Sample Average > B-Basis, Sample Minimum > A-basis, Coef. Of Var < 7 3.Sample Minimum > B-Basis 4.Sample Average > Sample Minimum > B-Basis 5.Sample Average > Sample Minimum > 6.Sample Average > Sample Minimum > 7.Sample Average > 8.Sample Average > Sample Minimum > Where and s are the mean and standard deviation of the qualifying sample k is the one-sided tolerance factor based on the number of test values and n is the number of tests in the equivalence sample

Agate Criteria –DOT/FAA/AR-03/19 specifies that the equivalency sample mean must be greater than: and the minimum value in the equivalency sample must be greater than: with: the qualifying sample mean the qualifying sample standard deviation n the size of the equivalency sample The k min and k mean values come from tables 20 and 21 in DOT/FAA/AR-03/19

C pk Criteria –A minimum Cpk of 1.33 is desired. –The equivalency sample is compared to test values are computed as follows: 1.The equivalency sample mean must be greater than: 2.The minimum value of the equivalency sample must be greater than: is the qualifying sample mean is the qualifying sample standard deviation

Comparison of Agate and C pk criteria Agate Criteria α set at 0.01 C pk Criteria C pk = 1.33 Mean Minimum

Comparison of Agate and C pk criteria For n = 5 and producers risk (α) = 0.01 k mean = ≈ C pk = k min.ind. = ≈ C pk = By allowing for a single retest of samples that fail the Agate criteria, the producers risk (α) drops to = For n = 5, k mean = ≈ C pk = k min.ind. = ≈ C pk = Agate criteria with one retest ≈ C pk criteria with no retest.

Type I and Type II Error Otherwise known as Producer’s risk and Consumer’s risk Claim that material is “good” Claim that material is “bad” Material is actually “good” No Error Type I Error (α) Producer’s Risk Material is actually “bad” Type II Error (β) Consumer’s Risk No Error

Comparison of Criteria for producer’s risk Criteria #2, #3, #4, #5, #6, and #8 all reject far too many ‘good’ samples. The producers risk for those samples range from approx. 10% at the lowest to nearly 90% at the highest. To simplify the presentation, the remaining slides examine only criteria #1, #7, Cpk and Agate.

Remaining Criteria CriteriaSample MeanSample Minimum # 1B-BasisA-Basis # 7 C pk Agate

Effect of the size of the Equivalency Sample What effect does the size of the equivalency sample have on the probability of rejection of ‘good’ samples?

Rejection rate of simulated “Good” samples Qualifying Sample Size 30

What about the consumer’s risk? How well do these criteria identify ‘Bad’ Samples? To assess the impact of various types of problems, the mean and the coefficient of variation of the simulated samples were changed to determine acceptance rates for the different criteria.

Samples were simulated to have the following problems: 1.The mean lowered from 100 to 95 2.The mean lowered by 1 std. dev. (σ) 3.The standard deviation (σ) was increased by 50% and the mean lowered by ½ σ. 4.Mean shifted randomly by 1 std dev. (σ) 5.Co-efficient of variation increased from 2 to 6 and to 10.

What should happen when the mean changes? If mean shifts down, then the acceptance rate should go down If mean shifts randomly (both up and down) then the variance is increased and the acceptance rate should go down.

Lowering the mean to 95

Lowering the mean by one Std. Dev.

Increasing the coefficient of variation and lowering the mean

What should happen when the coefficient of variation increases for the equivalency sample? When the equivalence sample has a larger coefficient of variation (or standard deviation) than the qualifying sample, the rejection rates should go up.

Rejection rate of simulated samples When C.V. is raised

Rejection rate of simulated samples When mean shifts randomly

What happens when the mean of the qualifying sample changes? The effect of shifting the mean of the qualifying sample up to the 95 th percentile or down to the 5 th percentile varies with the criteria being evaluated. Generally, as expected, a lower qualifying sample mean leads to a lower failure rate and vice versa For samples tested with the qualifying mean at the 95 th percentile (too high), the size of the equivalency sample impacted the failure rate.

Rejection rate of simulated samples When Qualifying mean changes

Effect of the size of the Equivalency Sample On almost all of the criteria tested, the size of the equivalency sample did have an effect. As sample size increases, so does the failure rate. This is due to the fact that all the other criteria involve checking the minimum value for the sample. As the sample size increases, the expected value for the sample minimum decreases. Unless the minimum value threshold is computed taking this into account (as the Agate criteria does), the probability of failure goes up with the sample size being evaluated. The one exception to this pattern was criteria #7, which only examined the mean, not the minimum value of the sample. In this case, as sample size increases, the sample mean converges to the true population mean and the failure rate falls.