Download presentation
Presentation is loading. Please wait.
Published byDamian Griffin Modified over 9 years ago
1
- 1 - Summary of P-box Probability bound analysis (PBA) PBA can be implemented by nested Monte Carlo simulation. –Generate CDF for different instances of the epistemic uncertainty –P-ox is given by bounds of probability distribution. –As a result, probability is interval valued quantity instead of single value.
2
- 2 - Validation metric in case of aleatory uncertainty Continuous model response and discrete experimental observations –Cumulative distribution function (CDF) from continuous PDF –Empirical distribution function (EDF) from finite number of data Example –Single data x=1.5 –3 data x=1.2 1.5 3.3 –100 data x= normrnd(3.5,1.2,1,100) dx=0.2; xx=0:dx:8; nx=length(xx); n=100; xp=normrnd(3.5,1.2,1,n); xp=sort(xp); %xp=[1.2 1.5 3.3] for i=1:nx; yy(i)=sum(heaviside(xx(i)-xp))/n; end stairs(xx,yy,'LineWidth',2); ylim([0 1])
3
- 3 - How to compare the two, and what is the metric ? –Compare visually the shape of the two distributions. –Compare values of moments such as mean and variance. –Compare maximum difference of CDF, Kolmogorov-Smirnov distance: –In this Oberkampf and Roy, area metric is suggested. area between model distribution F and data distribution S n as the measure of mismatch between the model and the experiment.
4
- 4 - Validation metric is better than comparing moments Figures explain: –Even if we have agreement with mean, we can still have large area. –Even if we have agreement with both mean & stdev, we can still have the large area metric. –Only when the two agree closely at every point, we have small area.
5
- 5 - area metric is better than KS metric –In the 1 st figure below, according to the K-S metrics, the right one is better than the left. But from engineering viewpoint, the left one is better. –In the 2 nd figure below, both give same K-S metric of unity. Area metric does not.
6
- 6 - area metric applies even when both are discrete Sometimes, we also have only a few simulation output from the model due to the computational cost.
7
Vicente Romero’s metric and critique Area metric does not distinguish between left and middle cases. Prefers metric in real space (see notes page for source)
8
- 8 - P-box in case of epistemic uncertainty –Case (a): p-box for interval uncertainty, i.e., no aleatory, only pure epistemic uncertainty. Then SRQ will also be in an interval. In terms of probability, CDF = 0 below left. CDF = 1 above right. –Case (c): degenerate p-box, i.e., precise CDF is given when there is only aleatory uncertainty. This is of course typical of traditional probability distribution. –Case (b): p-box for a mixture of aleatory & epistemic uncertainty. As a result, we get bounded CDF. e.g., at SRQ = 34, CDF is in the range [0.2, 0.6]. Compare with case (c) where the CDF is a single value.
9
- 9 - Area metric d under epistemic uncertainty –Consider only the case of single observation. –For easier understanding, think about the extreme case of aleatory uncertainty, in which the lower bound and upper bound is identical.
10
- 10 - Homework Consider the model used in previous lecture that simulates a response y(x) as a function of input x. At x=3.0, we have experimentally measured the performance of y three times to get 5.7, 5.84, 6.45. 1.Determine the confidence interval of the unknown mean of the difference between the model and experimental data. Plot the sample mean and its confidence interval. Based on the result, is the model is valid ? 2.Now the input x is found to be random variable, following normal distribution with the mean at 3, standard deviation 0.5. Due to this, the performance y is no longer deterministic but is given by the distribution. Use the crude Monte Carlo simulation to obtain samples of y and plot the histogram. From the samples, calculate the mean and standard deviation of output y. How much is the difference between the mean of y and deterministic output y ? 3.Plot empricial cumulative distribution function from experimental data as in slide 2 of this lecture. 4.Plot the two cdf's, one being the cdf of model, the other the empirical cdf in one figure. 5.Calculate the validation metric as defined by Kolmogorov-Smirnov test, which is the maximum difference of the two CDF's. 6.Calculate the validation metric as defined by area between two CDF as given by page 9.
11
- 11 - Homework -continued 7.Draw the Romero real validation metric based on the confidence intervals. 8.Now the model is changed to the following equation, which includes a parameter u, instead of a fixed value 2. and the value of u is known to be between (1,3). Assuming this as uniformly distributed, by employing double loop Monte-Carlo process, draw p-box of the cdf. 9.Plot empricial cumulative distribution function on top of the p-box. 10.Calculate the validation metric as defined by area between the two CDFs The terminology used in this lecture is very important!
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.