Sensitivity Analysis Reference n Bayesian Networks and Decision Graphs Finn V. Jensen n Expert Systems and Probabilistic Network Models Enrique Castillo,

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

CH-4 Ontologies, Querying and Data Integration. Introduction to RDF(S) RDF stands for Resource Description Framework. RDF is a standard for describing.
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
Rulebase Expert System and Uncertainty. Rule-based ES Rules as a knowledge representation technique Type of rules :- relation, recommendation, directive,
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Multisource Fusion for Opportunistic Detection and Probabilistic Assessment of Homeland Terrorist Threats Kathryn Blackmond Laskey & Tod S. Levitt presented.
An Introduction to Variational Methods for Graphical Models.
Introduction of Probabilistic Reasoning and Bayesian Networks
Pearl’s Belief Propagation Algorithm Exact answers from tree-structured Bayesian networks Heavily based on slides by: Tomas Singliar,
Using the Semantic Web to Construct an Ontology- Based Repository for Software Patterns Scott Henninger Computer Science and Engineering University of.
Assignment I, part 1. Groups of three students. Specify one as group leader. group names to TA and me. Create an object-oriented conceptualization.
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Dynamic Ontologies on the Web Jeff Heflin, James Hendler.
Learning with Bayesian Networks David Heckerman Presented by Colin Rickert.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
A Probabilistic Model for Classification of Multiple-Record Web Documents June Tang Yiu-Kai Ng.
. Bayesian Networks Lecture 9 Edited from Nir Friedman’s slides by Dan Geiger from Nir Friedman’s slides.
Logical and Probabilistic Reasoning to Support Information Analysis in Uncertain Domains Marco Valtorta, John Byrnes, and Michael Huhns
MISC.. 2 Integration Formula for consumer surplus Income stream - revenue enters as a stream - take integral of income stream to get total revenue.
1 Department of Computer Science and Engineering, University of South Carolina Issues for Discussion and Work Jan 2007  Choose meeting time.
July 3, A36 Theory of Statistics Course within the Master’s program in Statistics and Data mining Fall semester 2011.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graphs.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering Conflicts in Bayesian Networks January 23, 2007 Marco Valtorta
. PGM 2002/3 – Tirgul6 Approximate Inference: Sampling.
Statistics for Managers Using Microsoft® Excel 5th Edition
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Additional Slides on Bayesian Statistics for STA 101 Prof. Jerry Reiter Fall 2008.
Evidence and scenario sensitivities in naïve Bayesian classifiers Presented by Marwan Kandela & Rejin James 1 Silja Renooij, Linda C. van der Gaag, "Evidence.
More About Significance Tests
Virtual COMSATS Inferential Statistics Lecture-6
Statistical Decision Theory
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Mid-Term Review Final Review Statistical for Business (1)(2)
Lecture 5a: Bayes’ Rule Class web site: DEA in Bioinformatics: Statistics Module Box 1Box 2Box 3.
Statistical Sampling & Analysis of Sample Data
ECE 8443 – Pattern Recognition Objectives: Error Bounds Complexity Theory PAC Learning PAC Bound Margin Classifiers Resources: D.M.: Simplified PAC-Bayes.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.
1 BA 275 Quantitative Business Methods Confidence Interval Estimation Estimating the Population Proportion Hypothesis Testing Elements of a Test Concept.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Week 71 Hypothesis Testing Suppose that we want to assess the evidence in the observed data, concerning the hypothesis. There are two approaches to assessing.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Classification Techniques: Bayesian Classification
Uncertainty Management in Rule-based Expert Systems
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Berendt: Advanced databases, winter term 2007/08, 1 Advanced databases – Inferring implicit/new.
Probabilistic Reasoning Inference and Relational Bayesian Networks.
Sampling and Sampling Distributions. Sampling Distribution Basics Sample statistics (the mean and standard deviation are examples) vary from sample to.
Prediction of Soil Corrosivity Index: A Bayesian Belief Network Approach Gizachew A. Demissie, PhD student Solomon Tesfamariam,
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
From natural language to Bayesian Networks (and back)
Approximate Inference
Data Mining Lecture 11.
Classification Techniques: Bayesian Classification
More about Posterior Distributions
The implementation of a more efficient way of collecting data
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Parametric Methods Berlin Chen, 2005 References:
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Presentation transcript:

Sensitivity Analysis Reference n Bayesian Networks and Decision Graphs Finn V. Jensen n Expert Systems and Probabilistic Network Models Enrique Castillo, Jose Manuel Gutierrez, and Ali D. Hadi n Omniseer Project

Sensitivity Analysis Given a Bayesian network and evidence e, and some hypotheses n Sensitivity to evidence n Sensitivity to parameter

Sensitivity to evidence n Which evidence is in favor of /against/irrelevant for n Which evidence discriminate from ?

Sensitivity to evidence How to measure the sensitivity n Normalized likelihoods n Bayes factors n Fraction of achieved probability

Sensitivity to evidence Definition Let e be evidence and h a hypothesis. Suppose that we want to investigate how sensitive the result p(h|e) is to the particular set e. We say that evidence is sufficient if p(h|e’) is almost equal to p(h|e’). We then also say that e\e’ is redundant. The term almost equal can be made precise by selecting a threshold and requiring that. Note that is the fraction between the two likelihood ratios. e’ is minimal sufficient if it is sufficient, but no proper subset of e’ is so. e’ is crucial if it is a subset of any sufficient set. e’ is important if the probability of h changes too much without it. To be more precise, if, where is some threshold.

Sensitivity to parameters n how much the posterior probability of some event of interest changes with respect to the value of some parameter in the Bayesian network n We assume that the event of interest is the value of a target variable. The parameter is either a conditional probability or an unconditional prior probability

Sensitivity to parameters Theorem and Corollaries n Theorm 1: Let BN be a Bayesian network over the universe U. Let t be a parameter and let e be evidence entered in BN. Then, assuming proportional scaling, we have n Proof: The probability of an instantiation (x1,…,xn) is n Note that all the parameters appearing in the above product are associated with different variables, and some of them may be specified numerically. Thus p(x1,…,xn) is a monomial of degree less than or equal to the number of symbolic nodes.

Sensitivity to parameters Theorem and Corollaries n Corollary 1: Let BN be a Bayesian network over the universe U. Let t be a set of parameter for different distributions, and let e be evidence entered into BN. Then, assuming proportional scaling, P(e)(t) is a multi-linear polynomial over t n Proof: let t=(x,y). From the previous theorem, we have n If we have more than two parameters, we let t=(x,y), where y is a set of parameters. And repeat the arguments above.

Sensitivity to parameters Theorem and Corollaries n Corollary 2: Let BN be a Bayesian network over the universe U. Let t be a set of parameters for different distributions. Let a be a state of and let e be evidence. Then P(a|e)(t) is a fraction of two multi-linear polynomials over t. n Proof: Corollary 1 and fundamental rule

Sensitivity to parameters One-way sensitivity analysis n Let t be a parameter for BN and let e be evidence. Let a be a state of the target node. In one-way sensitivity analysis, we wish to determine p(e) and p(a,e) as functions of t. n Let t0 be the initial value of t. n Let t1 be the second value of t n Combing Corollary 2, we have

Sensitivity Analysis in Our Project n Project Introduction

Value of Information Sensitivity Analyzer Surprise Detector Bayesian Reasoning Service Project Overview BN Fragments Matcher Composer Instantiated Fragments Situation Specific Scenarios Tagged messages Modified Text John Doe London … John Doe … Bayesian Networks Documents Messages Events Tasks Massive Data

Bayesian Network Fragment Matching Example 1) Report Date: 1 April, FBI: Abdul Ramazi is the owner of the Select Gourmet Foods shop in Springfield Mall. Springfield, VA. (Phone number ). First Union National Bank lists Select Gourmet Foods as holding account number Six checks totaling $35,000 have been deposited in this account in the past four months and are recorded as having been drawn on accounts at the Pyramid Bank of Cairo, Egypt and the Central Bank of Dubai, United Arab Emirates. Both of these banks have just been listed as possible conduits in money laundering schemes. Partially- Instantiated Bayesian Network Fragment ….. …. BN Fragment Repository

Bayesian Network Fragment Composition Example Fragments Situation-Specific Scenario

Protégé overview n What is Protégé ? A tool which allows the user to: construct a domain ontology customize data entry forms enter data

OpenCyc overview What is OpenCyc ? o The open source version of the Cyc technologyCyc o World's largest and most complete general knowledge base and commonsense reasoning engine

OpenCyc overview --- cont. Where can we use OpenCyc ? o speech understanding o database integration o rapid development of an ontology in a vertical area o prioritizing, routing, summarization, and annotating

OpenCyc overview --- cont. What does OpenCyc look like ?

OpenCyc overview --- cont. More Detail Here

RDF overview What is RDF? n Stands for Resource Description Framework n Recommended by the World Wide Web Consortium (W3C) n Model meta-data about the resources of the web

RDF overview --- Cont. What does RDF file look like? Basically, there are two kinds of file in RDF system n RDFS file --- The schema file n RDF file --- The file containing all instances

RDF overview --- Cont. RDFS file <!DOCTYPE rdf:RDF [ ]>

RDF overview --- Cont. RDF file

Protégé GUI—Class Design

Protégé GUI—Instance View

Sensitivity Analysis in Our Project n Sensitivity analysis assesses how much the posterior probability of some event of interest changes with respect to the value of some parameter in the model n We assume that the event of interest is the value of a target variable. The parameter is either a conditional probability or an unconditional prior probability n If the sensitivity of the target variable having a particular value is low, then the analyst can be confident in the results, even if the analyst is not very confident in the precise value of the parameter n If the sensitivity of the target variable to a parameter is very high, it is necessary to inform the analyst of the need to qualify the conclusion reached or to expend more resources to become more confident in the exact value of the parameter

Example: Case Study #4 Computing Sensitivity 2

Example: Case Study #4 Computing Sensitivity In the context of the information already acquired, i.e., travel to dangerous places, large transfers of money, etc., the parameter that links financial irregularities to being a suspect is much more important for assessing the belief in Ramazi being a terrorist than the parameter that links dangerous travel to being a suspect. The analyst may want to concentrate on assessing the first parameter precisely.

Sensitivity Analysis: Formal Definition n Let the evidence be a set of findings: n Let t be a parameter in the situation-specific scenario n Then, [Castillo et al., 1997; Jensen, 2000] n α and β can be determined by computing P(e) for two values of t n More generally, if t is a set of parameters, then P(e)(t) is a linear function in each parameter in t, i.e., it is a multi-linear function of t n Recall that n Then, n We can therefore compute the sensitivity of a target variable V to a parameter t by repeating the same computation with two values for the evidence set, viz. e and

Algorithm and Implementation n Bucket Elimination n Goal-oriented Symbolic propagation n Differential Approach to Inference in BN A Differential Approach to Inference in Bayesian Networks Adnan Darwiche n A Computational Architecture for N-way Sensitivity Analysis of Bayesian Networks Veerle M. H. Coupé, Finn V. Jensen, Uffe Kjærulff & Linda C. van der Gaag