Edoardo Tosonia,b, Ahti Saloa, Enrico Ziob,c

Slides:



Advertisements
Similar presentations
Introduction to Transportation Systems. PART II: FREIGHT TRANSPORTATION.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
SA-1 Probabilistic Robotics Planning and Control: Partially Observable Markov Decision Processes.
CSE-573 Artificial Intelligence Partially-Observable MDPS (POMDPs)
Engineering Economic Analysis Canadian Edition
Planning under Uncertainty
POMDPs: Partially Observable Markov Decision Processes Advanced AI
Optimizing General Compiler Optimization M. Haneda, P.M.W. Knijnenburg, and H.A.G. Wijshoff.
Approximate Dynamic Programming for High-Dimensional Asset Allocation Ohio State April 16, 2004 Warren Powell CASTLE Laboratory Princeton University
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Computational Stochastic Optimization: Bridging communities October 25, 2012 Warren Powell CASTLE Laboratory Princeton University
1 Performance Evaluation of Computer Networks: Part II Objectives r Simulation Modeling r Classification of Simulation Modeling r Discrete-Event Simulation.
Engineering Economic Analysis Canadian Edition
Yaomin Jin Design of Experiments Morris Method.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
The Birthday Problem Section The Birthday Problem What is the probability that in a group of 37 students at least two of them share a birthday?
Simulation is the process of studying the behavior of a real system by using a model that replicates the behavior of the system under different scenarios.
Uncertainty Management in Rule-based Expert Systems
ECE 466/658: Performance Evaluation and Simulation Introduction Instructor: Christos Panayiotou.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Maximum Flow Problem (Thanks to Jim Orlin & MIT OCW)
15.082J and 6.855J March 4, 2003 Introduction to Maximum Flows.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
Best in Market Pricing. What is Best in Market Pricing ? An extension of parametric modeling for negotiating lowest pricing for a statement of work consisting.
Basic Practice of Statistics - 3rd Edition Introducing Probability
Edoardo Tosonia,b, Ahti Saloa, Enrico Ziob,c
P & NP.
Algorithms and Problem Solving
Chapter 7. Classification and Prediction
Market-Risk Measurement
Control Flow Testing Handouts
Analysis of Algorithms
CompSci 230 Software Construction
Making complex decisions
Outline of the Chapter Basic Idea Outline of Control Flow Testing
CH. 2: Supervised Learning
Future Studies (Futurology)
Aditya P. Mathur Purdue University
Classification with Perceptrons Reading:
Data Mining (and machine learning)
Data Mining Lecture 11.
Multi - Way Number Partitioning
Physics 114: Lecture 10 Error Analysis/ Propagation of Errors
Data Mining Practical Machine Learning Tools and Techniques
Risk-informed Decision Making under Incomplete Information
Routing and Logistics Arc Routing 2018/11/19.
Hidden Markov Models Part 2: Algorithms
Algorithm An algorithm is a finite set of steps required to solve a problem. An algorithm must have following properties: Input: An algorithm must have.
TransCAD Vehicle Routing 2018/11/29.
A. Mancusoa,b, M. Compareb, A. Saloa, E. Ziob,c
Introduction to Maximum Flows
Lecture 2 – Monte Carlo method in finance
of the Artificial Neural Networks.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Propagation Algorithm in Bayesian Networks
Basic Practice of Statistics - 3rd Edition Introducing Probability
Matching Methods & Propensity Scores
Recurrences (Method 4) Alexandra Stefan.
Basic Practice of Statistics - 3rd Edition Introducing Probability
Algorithms and Problem Solving
Ensemble learning.
D1 Discrete Mathematics
This Lecture Substitution model
Essential Statistics Introducing Probability
Reinforcement Learning Dealing with Partial Observability
Independence and Counting
Error Correction Coding
Basic Practice of Statistics - 5th Edition Introducing Probability
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Edoardo Tosonia,b, Ahti Saloa, Enrico Ziob,c Bayesian networks for comprehensive scenario analysis of nuclear waste repositories Edoardo Tosonia,b, Ahti Saloa, Enrico Ziob,c Systems Analysis Laboratory, Dept of Mathematics and Systems Analysis - Aalto University Laboratory of signal and risk analysis, Dipartimento di Energia - Politecnico di Milano Chair on Systems Science and the Energetic Challenge - École Centrale Paris and Supelec June 21, 2017

? Scenario analysis of nuclear waste repository Scenario Analysis Deep geological disposal of nuclear waste Performance Assessment: Many physical and chemical phenomena interacting Long time horizon (10,000 – 1,000,000 years) Large aleatory uncertainty about the evolution of the disposal system Scenario 1 ? Scenario Analysis Safety Scenario 2 Scenario ... Scenario n

Challenges in scenario analysis Review of methodologies for Scenario Analysis of nuclear waste repositories Very few repositories licensed to date Challenges: Bayesian network (BN) of the disposal system 1 Building a system model to guide scenario development Algorithm to select which simulations to run for analyzing scenarios 2 Ensuring comprehensiveness Conservative aggregation of multiple experts’ beliefs on probabilities in the BN 3 Treating the epistemic uncertainties

Definitions FEPs Set of nodes and of directed arcs Stochastic variables with discrete states e.g. Probabilities Independent nodes → Unconditional Dependent nodes → Conditional Radiological consequence Scenarios? Combination of FEPs states 7 FEPs 3 states per FEP 37 = 2,187 scenarios

Bayesian network - safety State of unacceptable consequences Each scenario z: Occurs with probability Causes failure with probability Overall failure probability of the disposal system Failure! Sum over all scenarios Probability of the scenario z Failure probability in the scenario z Safety: How to obtain these probabilities?

EPISTEMIC UNCERTAINTY Expert judgments Set of experts at each node Experts’ beliefs on the probabilities EPISTEMIC UNCERTAINTY For each probablity multiple values e.g. Weight to expert e at node i

Optimization What values of the weights? Assign values to obtain the most conservative estimate of the failure probability, given the elicited beliefs Optimization model:

Very strong assumption! Deterministic tools in the BN (1/2) Some dependences modeled by deterministic tools (computer codes, laboratory experiments) Better capure the physical phenomena Resource-intensive simulations (time, cost...) qNF qB Very strong assumption! Soon removed Another option for obtaining the probabilities is to resort to deterministic tools. Specifically, some dependences in the Bayesian network can be modeled by deterministic tools like computer codes or laboratory experiments. In our example, we modeled the dependence of the Radionuclide discharge on its parents Canister breach, Buffer flow and Near-field flow by a computer code of radionuclide transport. In general, when such tools are available, they may be preferred to expert judgment, as they are supposed to better capture the physical phenomena behind the dependences. Another implication of using a deterministic tool is that we get a conditional probability table with integer conditional probabilities. Specifically, each row corresponds to a computer-code simulation whose input is a specific combination of states of the parent nodes, and whose output is a deterministic prediction of the state of Radionuclide discharge. This is actually a very strong assumption, which we are going to remove soon from our work. The implication is that we will have to solve the problem differently form what you will see in the next slides, but the underlying idea will stay the same, and this is why i will still present the next slides. Anyway, at the beginning of the analysis, the table is empty, and we can fill it in by progressively simulating all the rows. If we can afford running all simulations, we should just do that. However, the problem is that each simulation can be resource-intensive in terms of computational time and also cost, especially in the case of lab experiments. So, we may not be able or not be willing to run simulations for all the rows, which are 27 in this case. D r

Deterministic tools in the BN (2/2) PFail depends on the realization of the table Finite discrete set of table realizations MONOTONICITY! Here, at the beginning: The interval may not be conclusive for assessing safety, if As long as there are empty rows in the table, the failure probability is not uniquely determined, even more so before any simulation is run and the table is totally empty. If we look at the equation of the failure probability, even if we assume that we have all the other probabilities in the network, for instance by expert judgment, we’re still missing the probabilities for the last node, because they’re missing from the table. In practice, the failure probability depends on what will be the realization of the table. But how many ex ante possible realizations of the table are there? Well, because of the strong assumption of integer probabilities, there is a finite set of them. The table might look like this, like this, like this. Proceeding in this way, the number of possible realizations would be almost intractable. But if we introduce resonable assumptions of monotonocity, this number drop down considerably. Anyway, we compile the full collection of ex ante possible realizations, and compute the failure probability corresponding to each of them. Then we take the smallest and the largest of these values to build an interval for the failure probability. This interval may not be conclusive for safety. An interval is conclusive if it is entirely below the maximum threshold epsilon for the failure probability, in what case the system is safe, or if it is entirely above the threshold, in what case the system is unsafe. In our example, when the table was still empty, the failure probability interval was 0 to 71%, which was not conclusive because we had set the threshold epsilon at 5%. Then the idea is, if we run one more simulation, we can fill in the corresponding row of the table, we can discard realization which are not possible anymore in the light of the new information, and we update the failure probability interval. And if the updated interval is still non conclusive, we iterate. Here comes the challenge of comprehensiveness in our Bayesian network: instead of running all simulations, can we simulate a restricted number of rows, corresponding to a restricted number of scenarios, and still be able to obtain a failure-probability interval which is conclusive for safety? Run simulation → Fill in the corresponding row → Discard realizations → Update interval COMPREHENSIVENESS

Simulation-selection algorithm Algorithm for iteratively selecting next simulation Assign a conclusivity score to each simulation Run the simulation with the highest score In order to answer this question, we’re in the process of developing an algorithm. The machinery of the algorithm will change completely after we will remove the assumption of integer probabilities. The underlying idea will stay the same, and I will now try to communicate it. At each iteration, the algorithm selects the next simulation to be run, in the attempt of obtaining a conclusive interval for the failure probability with the least number of simulations. Basically we wish to give a conclusivity score to the remaining simulations. For each simulation, we anticipate its possible results, and we anticipate the failure-probability interval as it would be updated after each of the possible results. Then we score each interval by two attributes which try to capture how conclusive the interval is. By averaging over the possible intervals, we obtain the conclusivity score for the simulation. We do the same for all simulations, and we run the simulation with the highest conclusivity score. 0.31 1 1 0.155 Simulation’s score: 0.78

Selected simulations [ 0 - 0.709] [ 0.016 - 0.709] [ 0.035 - 0.709] 1 [ 0 - 0.709] 1 1 [ 0.016 - 0.709] [ 0.035 - 0.709] 1 [ 0.035 - 0.090] [ 0.035 - 0.077] [ 0.035 - 0.060] [ 0.035 - 0.046] Here, we briefly show the results of the algorithm. First it tells us to simulate the last row, we run it, and we update the interval, which is still very large. We proceed, and at the third simulation the interval gets much narrower, but still across the maximum acceptable failure probability. Finally, at the sixth simulation we obtain a failure-probability interval which is conclusive because it is entirely below the threshold, whereby we can say that the system is safe. In summary, instead of simulating all 27 rows, we ran a restricted set of 6 simulations which was comprehensive because we obtained a conclusive interval for the failure probability. 1 1

Summary and next steps Methodology to address challenges in Scenario Analysis of nuclear waste repository Just a few notes on the next steps for improving the methodology: let me just emphasize the attachment of post-processing techniques for identifying the most important failure scenarios, and the removal of the assumptions of integer probabilities from computer codes, as i said many times.