An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 10 The Analysis of Variance.
Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
Markov Chain Monte Carlo Convergence Diagnostics: A Comparative Review By Mary Kathryn Cowles and Bradley P. Carlin Presented by Yuting Qi 12/01/2006.
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
Bayesian Methods with Monte Carlo Markov Chains III
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Markov processes in a problem of the Caspian sea level forecasting Mikhail V. Bolgov Water Problem Institute of Russian Academy of Sciences.
Visual Recognition Tutorial
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
MARE 250 Dr. Jason Turner Hypothesis Testing II To ASSUME is to make an… Four assumptions for t-test hypothesis testing: 1. Random Samples 2. Independent.
Chapter 6 Introduction to Sampling Distributions
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Evaluating Hypotheses
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
The Lognormal Distribution
Component Reliability Analysis
Introduction to Statistical Inferences
Simulation Output Analysis
AN ITERATIVE METHOD FOR MODEL PARAMETER IDENTIFICATION 4. DIFFERENTIAL EQUATION MODELS E.Dimitrova, Chr. Boyadjiev E.Dimitrova, Chr. Boyadjiev BULGARIAN.
The horseshoe estimator for sparse signals CARLOS M. CARVALHO NICHOLAS G. POLSON JAMES G. SCOTT Biometrika (2010) Presented by Eric Wang 10/14/2010.
Verification & Validation
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
10-1 Introduction 10-2 Inference for a Difference in Means of Two Normal Distributions, Variances Known Figure 10-1 Two independent populations.
1 Physical Fluctuomatics 5th and 6th Probabilistic information processing by Gaussian graphical model Kazuyuki Tanaka Graduate School of Information Sciences,
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Exam I review Understanding the meaning of the terminology we use. Quick calculations that indicate understanding of the basis of methods. Many of the.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Sequential Experimental Designs For Sensitivity Experiments NIST GLM Conference April 18-20, 2002 Joseph G. Voelkel Center for Quality and Applied Statistics.
Quantification of the non- parametric continuous BBNs with expert judgment Iwona Jagielska Msc. Applied Mathematics.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
1 SMU EMIS 7364 NTU TO-570-N Inferences About Process Quality Updated: 2/3/04 Statistical Quality Control Dr. Jerrell T. Stracener, SAE Fellow.
Lab 3b: Distribution of the mean
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
Problem: 1) Show that is a set of sufficient statistics 2) Being location and scale parameters, take as (improper) prior and show that inferences on ……
University of Ostrava Czech republic 26-31, March, 2012.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Bayesian Travel Time Reliability
Brief Review Probability and Statistics. Probability distributions Continuous distributions.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
CS246 Latent Dirichlet Analysis. LSI  LSI uses SVD to find the best rank-K approximation  The result is difficult to interpret especially with negative.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Sampling and Sampling Distributions. Sampling Distribution Basics Sample statistics (the mean and standard deviation are examples) vary from sample to.
Generating Random Variates
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
Rutgers Intelligent Transportation Systems (RITS) Laboratory
Statistics in Applied Science and Technology
Unfolding Problem: A Machine Learning Approach
Econ 3790: Business and Economics Statistics
Convergence of Sequential Monte Carlo Methods
Presentation transcript:

An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology

Sensitivity experiments: experimental settings in which each experimental unit has a critical stimulus level that cannot be observed directly. Introduction

Response Curve F(x) is the distribution of the critical stimulus levels over the test specimens. where, G is a known distribution function The general version of F(x) is

Observations stimulus settings x 1,…,x n, and corresponding response results y 1,…,y n. When x i is at or above the critical level of the i th unit, it responds and y i =1 ; otherwise, it does not respond and y i =0, i=1,…,n.

The interested parameter We are often interested in pth quantile of F(x),  p, Our goal Make inference for  p using small samples.

Historic Data In China, for explosives and materials, there often exists a data set from the documented method up-and-down procedure for analyzing the sensitivity. When considering the availability, interest would aim to estimate the extreme quantile  p by limited sample size n.

inappropriate for the existed methods Based on such kind of historic data set, we can use Wu’s method, Neyer’s method, or Joseph’s method to estimate  p. However, for this directly using, there will cast many units in the subsequent sequential procedure to modify the characteristic of centering on  0.5 of this historic data set before quickly converging to the interest parameter  p.

Proposed methods Our purpose is to develop a new sequential procedure that can make the estimate for  p more precisely by quickly and efficiently exploiting the information in the tested data and known knowledge.

Specification of the Prior Let  =  2 /  1,  =1/  1. Thus It is natural to assume that the prior distribution of  is normal, the prior distribution of  is lognormal, and they are independent.

For the historic data set, the experiment settings is a realization of the Markov chain with finite state space { a 0,…,a k-1 } and transition probability matrix where

Let be the MLE of ( ,  ) based on historic data set. Now, with as transition probability, generate m 1 Markov chain realizations, and obtain estimates of ( ,  ). Then with as transition probability, generate the m 2 second-level Markov chain realizations and obtain estimates

After the transformation from ( ,  ) to (  1,  2 ), we specify the prior distribution for (  1,  2 ). Let From we calculate their means and variances and then specify the prior distribution of  and .

Assume we have collected data (x 1, y 1 ),…, ( x t, y t ) (This includes the case of not having any data, for which t =0). The New Sequential Design Our objective: select a new level x* so that if we run there and obtain the result y x, the average posterior variance of  with respect to y x, but conditioned on (x 1, y 1 ),…,(x t, y t ), E n [Var(  |y 1,…,y t,y x )], is minimized. Set x t+1 =x*, make a run there and obtain the observation y t+1.

Estimation of  After a certain fixed number, n, of observations is obtained, we use the posterior expectation of  to estimate it. Let  be the parameter space for  =(  1,  2 ) and let x i (i  1) take values in A. Let the subset of x  1 –  2 for x  A,  be B. Some notations

High Order Approximation Theorem 1. Suppose the following conditions hold: (i) There exists an integer m such that the data set {(x i, y i ): i=1,…,m} has an overlap in responses and nonresponses. (ii) -lnG(u) and –ln[1-G(u)] are convex. G(u) is strictly increasing at every u satisfying 0<G(u)<1. (iii) The derivatives of five orders of G(u) are bounded over {u: 0<G(u)<1}. (iv) A is bounded and B  {u: 0<G(u)<1}. Also, B is compact and connected. Then for n  m Laplace formula produces the approximation

Consistency of the posterior estimator Theorem 2 Under conditions of Theorem 1 and following condition is consistent. B 1 ={(u,v): u=1, |v|<  1 <1}, B 2 ={(u,v): u  1, u>  2 >0,v=0}, B 3 ={(u,v): u  1, v  0, u>  31 >0, h(u,v)>  32 >0}

Comparisons in Simulation We now compare the performance of our method with other four methods through simulations. Under comparison are (a) the up and down method recommended by ASTM (1995); (b) Wu’s recursive design (1985); (c) Neyer’s method (1994); (d) the modified Robbins-Monro procedure proposed by Joseph (2004).

We use the logit model, as the true distribution, where  1 =1/2,  2 =5. We first simulate the historic data set with sample size 20 recommended by standards. Then simulate the data by using our method and other four methods with sample size n. We repeat above processes 500 times and obtain 500 estimates of .

Simulating Results The results of Monte Carlo mean and mean squared error of 500 estimates are given in following table.  0.1 (true value is 5.61)  0.2 (true value is 7.23) Design nn Mean MSE Proposed method Wu’s method Neyer’s method Roshan’s method Up-and–down method

The simulation study demonstrates the superior performance of our method over the other methods. Our method did a good job for commonly used extreme quantiles under the requested sample size. Conclution

Thank you for your attendance.

All involved integrals are two dimensional integrals, In this paper, we use the algorithm which is efficient and easy to implement given by James C. Fu and Liquan Wang (2002) to calculate this kind of integral. Here we don’t use MCMC algorithms because for the low dimensional integral MCMC does not perform very well and the functional forms of full conditionals of posterior distributions is not clear. are not very easy to be obtained.

Voelkel J. G.