Semantically-Linked Bayesian Networks: A Framework for Probabilistic Inference Over Multiple Bayesian Networks PhD Dissertation Defense Advisor: Dr. Yun.

Slides:



Advertisements
Similar presentations
Slide 1 of 18 Uncertainty Representation and Reasoning with MEBN/PR-OWL Kathryn Blackmond Laskey Paulo C. G. da Costa The Volgenau School of Information.
Advertisements

Autonomic Scaling of Cloud Computing Resources
Big Ideas in Cmput366. Search Blind Search State space representation Iterative deepening Heuristic Search A*, f(n)=g(n)+h(n), admissible heuristics Local.
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Exact Inference. Inference Basic task for inference: – Compute a posterior distribution for some query variables given some observed evidence – Sum out.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
Knowledge Representation and Reasoning University "Politehnica" of Bucharest Department of Computer Science Fall 2010 Adina Magda Florea
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
UMBC an Honors University in Maryland 1 Uncertainty in Ontology Mapping: Uncertainty in Ontology Mapping: A Bayesian Perspective Yun Peng, Zhongli Ding,
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Introduction of Probabilistic Reasoning and Bayesian Networks
Probabilistic Reasoning with Uncertain Data Yun Peng and Zhongli Ding, Rong Pan, Shenyong Zhang.
Causal and Bayesian Network (Chapter 2) Book: Bayesian Networks and Decision Graphs Author: Finn V. Jensen, Thomas D. Nielsen CSE 655 Probabilistic Reasoning.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
On the Role of MSBN to Cooperative Multiagent Systems By Y. Xiang and V. Lesser Presented by: Jingshan Huang and Sharon Xi.
A Probabilistic Framework for Information Integration and Retrieval on the Semantic Web by Livia Predoiu, Heiner Stuckenschmidt Institute of Computer Science,
PR-OWL: A Framework for Probabilistic Ontologies by Paulo C. G. COSTA, Kathryn B. LASKEY George Mason University presented by Thomas Packer 1PR-OWL.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.6 [P]: Reasoning Under Uncertainty Sections.
Semantics For the Semantic Web: The Implicit, the Formal and The Powerful Amit Sheth, Cartic Ramakrishnan, Christopher Thomas CS751 Spring 2005 Presenter:
1 Department of Computer Science and Engineering, University of South Carolina Issues for Discussion and Work Jan 2007  Choose meeting time.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graphs.
Causal Models, Learning Algorithms and their Application to Performance Modeling Jan Lemeire Parallel Systems lab November 15 th 2006.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Aspects of Bayesian Inference and Statistical Disclosure Control in Python Duncan Smith Confidentiality and Privacy Group CCSR University of Manchester.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
A Brief Introduction to Graphical Models
Ontology Development Kenneth Baclawski Northeastern University Harvard Medical School.
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Reasoning in Uncertain Situations
Introduction to Bayesian Networks
Generalizing Variable Elimination in Bayesian Networks 서울 시립대학원 전자 전기 컴퓨터 공학과 G 박민규.
Bayesian Nets and Applications. Naïve Bayes 2  What happens if we have more than one piece of evidence?  If we can assume conditional independence 
Physics Fluctuomatics (Tohoku University) 1 Physical Fluctuomatics 12th Bayesian network and belief propagation in statistical inference Kazuyuki Tanaka.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Announcements Project 4: Ghostbusters Homework 7
Probabilistic Networks Chapter 14 of Dechter’s CP textbook Speaker: Daniel Geschwender April 1, 2013 April 1&3, 2013DanielG--Probabilistic Networks1.
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
On the Role of Multiply Sectioned Bayesian Networks to Cooperative Multiagent Systems Presented By: Yasser EL-Manzalawy.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
A Bayesian Perspective to Semantic Web – Uncertainty modeling in OWL Jyotishman Pathak 04/28/2005.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Introduction on Graphic Models
Probabilistic Reasoning Inference and Relational Bayesian Networks.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Integrative Genomics I BME 230. Probabilistic Networks Incorporate uncertainty explicitly Capture sparseness of wiring Incorporate multiple kinds of data.
Graduate School of Information Sciences, Tohoku University
Presented By S.Yamuna AP/CSE
Jie Bao, Doina Caragea and Vasant G Honavar
Inconsistent Constraints
CAP 5636 – Advanced Artificial Intelligence
CS 188: Artificial Intelligence
Graduate School of Information Sciences, Tohoku University
Graduate School of Information Sciences, Tohoku University
Probabilistic Reasoning
Graduate School of Information Sciences, Tohoku University
Read R&N Ch Next lecture: Read R&N
Graduate School of Information Sciences, Tohoku University
Presentation transcript:

Semantically-Linked Bayesian Networks: A Framework for Probabilistic Inference Over Multiple Bayesian Networks PhD Dissertation Defense Advisor: Dr. Yun Peng Rong Pan Department of Computer Science and Electrical Engineering University of Maryland Baltimore County Aug 2, 2006

Outline ► Motivations ► Background ► Overview ► How Knowledge is Shared ► Inference on SLBN ► Concept Mapping using SLBN ► Future works

Motivations (1) ► Separately developed BNs about  related domains  different aspects of the same domain …

Motivations (2) ► Existing approach:  Multiply Sectioned Bayesian Networks (MSBN)  Every subnet is sectioned from a global BN  Strictly consistent subnets  Exactly identical shared variables with same distribution  All parents of the shared variables must appear in one subnet Sectioning

Motivations (3) ► Existing approach:  Agent Encapsulated Bayesian Networks (AEBN)  Distribution BN Model for a specific application  Hierarchical global structure  Very restricted expressiveness  Exactly identical shared variables with different prior distributions Agent Output Variable Input Variable Local Variable

Motivations (4) ► A distributed BN model was expected with features:  Uncertainty reasoning over separately developed BNs  Variables shared by different BNs can be similar but not identical  Principled, well justified  Support various applications

Background Bayesian Network ► DAG ► Variables  with Finite States ► Edges: causal influences ► Conditional Probability Table (CPT)

Background Evidences in BN Original BNHard Evidence: Male_Mammal = True Soft Evidence: Q(Male_Mammal) = ( ) Virtual Evidence: L(Male_Mammal) = 0.8/0.2 Virtual Evidence = Soft Evidence: L(Male_Mammal) = 0.3/0.2

Background Jeffrey’s Rule (Soft Evidence) ► Given external observations Q(B i ), the rest of the BN is updated by Jeffrey’s Rule: where P(A| B i ) is the conditional probability before evidence, Q(B i ) is the soft evidence. where P(A| B i ) is the conditional probability before evidence, Q(B i ) is the soft evidence. ► Multiple Soft Evidences  Problem: update one variable’s distribution to its target value can make those of others’ off their targets  Solution: IPFP

Background Iterative Proportional Fitting Procedure (IPFP) ► Q 0 : initial distribution on the set of variables X, ► {P(S i )}: a consistent set of n marginal probability distributions, where X  Si  . ► The IPFP process where i is the iteration number, j = (i-1) mod n + 1 ► The distribution after IPFP satisfies the given constraints {P(S i )} and has minimum cross-entropy to the initial distribution Q 0

SLBN: Overview (1) ► Semantically-Linked Bayesian Networks (SLBN)  A theoretical framework that supports probabilistic inference over separately developed BNs Global Knowledge Similar variables

SLBN: Overview (2) ► Features  Inference over separate BNs that share semantically similar variables  Global knowledge: J-graph  Principled, well-justified ► In SLBN  BNs are linked at the similar variables  Probabilistic influences are propagated via the shared variables  Inference process utilizes Soft Evidence (Jeffrey’s Rule), Virtual Evidence, IPFP, and traditional BN inference

How knowledge is shared: Semantic Similarity (1) What is similarity? Similar: Pronunciation: 'si-m&-l&r, 'sim-l&r Function: adjective 1: having characteristics in common 2: alike in substance or essentials 3: not differing in shape but only in size or position –– High-tech Company Employee V.S. High-income People Computer Keyboard V.S. Typewriter

How knowledge is shared: Semantic Similarity (2) ► Semantic Similarity of concepts  Share of common instances  Quantified and utilized with direction  Quantified by the ratio of the shared instances to all the instances  Natural language’s definition for “similar” is vague  Hard to formalize  Hard to quantify  Hard to utilize in intelligence Conditional Probability P(High-tech Company Employee | High-income People)

Man V.S. Woman How knowledge is shared: Variable Linkage (1) ► In Bayesian Network (BN) / SLBN  Concepts are represented by variables  Semantic similarities are between propositions We say “High-tech Company Employee” is similar to “High-income People” We mean “High-tech Company Employee = True” is similar to “High-income People = True”

How knowledge is shared: Variable Linkage (2) ► Variable linkages  Represent semantic similarities in SLBN  Are between variables in different BNs A : Source Variable B : Destination Variable N A : Source BN N B : Destination BN : Quantification of the similarity is a m × n matrix:

How knowledge is shared: Variable Linkage (3) ► Variable Linkage V.S. BN Edge Variable Linkage BN Edge Representation Of Semantic Similarity Causal Influences Conditional Probability Quantification of Similarity; Invariant w.r.t. any event Conditional dependency; May be changed by events Prob. Influence Propagation Along the direction Both directions ( -msg. and  -msg.) ( -msg. and  -msg.)

How knowledge is shared: Variable Linkage (4) ► Expressiveness of Variable Linkage  Logical relationships defined in OWL syntax: Equivalent, Union, Intersection, and Subclass complement.  Relaxation of logical relationships by replacing set inclusion by overlapping: Overlap, Superclass, Subclass  Equivalence relations but same concepts are modeled as different variables

How knowledge is shared: Examples (1) … … Identical Union

How knowledge is shared: Examples (2) Overlap Superclass …

How knowledge is shared: Consistent Linked Variables ► The priori beliefs on the linked variables on both sides must be consistent with the variable linkage:  P 2 (B) = ∑ i P S (B|A=a i )P 1 (A=a i )  There exists a single distribution consistent with the prior belief on A, B,  A,  B, and the linkage’s similarity. ► examined by IPFP AB AA BB P 1 (  A ) P 1 (A|  A ) P 1 (A) P 2 (  B ) P 2 (B|  A ) P 2 (B) P S (B| A)

Inference on SLBN The Process 1. Enter Evidence 2. Propagate 4. Updated Result 3. Enter Soft/Virtual Evidences; BN Belief Update With traditional Inference BN Belief Update With traditional Inference SLBN Rules for Probabilistic Influence Propagation SLBN Rules for Probabilistic Influence Propagation BN Belief Update with Soft Evidence BN Belief Update with Soft Evidence

Inference on SLBN The Theory Bayes’ RuleJeffrey’s RuleIPFP Soft Evidence BN Inference Virtual Evidence SLBN Theoretical Basis Implementation (Existing) Implementation (SLBN)

Inference on SLBN Assumptions/Restrictions ► All linked BNs are consistent with the linkages ► One variable can only be involved in one linkage ► Causal precedence in all linked BNs are consistent Linked BNs with inconsistent causal sequences Linked BNs with consistent causal sequences

Inference on SLBN Assumptions/Restrictions (Cont.) ► For a variable linkage, the causes/effects of source is also the causes/effects of the destination  Linkages cannot cross each other … … …... Crossed linkages

Inference on SLBN Inference on SLBN SLBN Rules for Probabilistic Influence Propagation (1) ► Some hard evidence influence the source from bottom Y1 Y1 Y2Y2 Y3 Y3 … … X1 X1 ► Propagated influences are represented by soft evidences ► Beliefs of destination BN are update with SE

Inference on SLBN Inference on SLBN SLBN Rules for Probabilistic Influence Propagation (2) ► Some hard evidence influence the source from top Y1 Y1 Y2Y2 Y3 Y3 … … X1 X1 ► Additional soft evidences are created to cancel the influences from the linkage to parent(dest(L))

Inference on SLBN Inference on SLBN SLBN Rules for Probabilistic Influence Propagation (3) ► Some hard evidence influence the source from both top and bottom Y1 Y1 Y2Y2 Y3 Y3 … … X1 X1 ► Additional soft evidences are created to propagate the combined influences from the linkage to parent(dest(L))

Inference on SLBN Belief Update with Soft Evidence (1) ► Represent soft evidences by virtual evidences  Belief update with soft evidence is IPFP  Belief update with one virtual evidence is one step of IPFP ► Therefore, we can  Use virtual evidence to implement IPFP on BN  Use virtual evidence to implement soft evidence ► SE VE  Iterate on the whole BN  Iterate on soft evidence variables

Inference on SLBN Belief Update with Soft Evidence (2) ► Iterate on whole BN AB (0.3, 0.7) (0.8, 0.2) (0.6, 0.4) (0.66, 0.34) (0.47, 0.53) (0.5, 0.5) (0.6, 0.4) (0.60, 0.40) (0.55, 0.45) (0.5, 0.5) …………. (0.6, 0.4) (0.5, 0.5) Q(B) = (0.5, 0.5) Q(A) = (0.6, 0.4) ve AB

Inference on SLBN Belief Update with Soft Evidence (1) ► Iterate on SE variables Q(B) = (0.5, 0.5)Q(A) = (0.6, 0.4) ve AB A = t A = f B = t B = f P(A, B) = IPFP with Q(A), Q(B) A = t A = f B = t B = f Q(A, B) =

Inference on SLBN Belief Update with Soft Evidence (3) ► Existing approaches : Big-Clique Big-Clique Iteration on whole BN Iteration on se variables BN Inference Basis RewriteJunction-Tree Wrapper of any method Time for each Iteration O(e |C| ) O(BN Inf.) O(e |V| ) Space O(e |C| ) O(|V|) O(e |V| ) C: the big clique V: se variables |C|≥|V| Iteration on whole BN: Small BNs, many soft evidences Iteration on se variables: Large BNs, a few soft evidences

J-Graph (1) Overview ► Joint-graph (J-graph) is a graphical probability model that represents  The joint distribution of SLBN  The interdependencies between variables across variable linkages ► Usage  Check if all assumptions are satisfied  Justify Inference Process

J-Graph (2) Definition ► J-Graph is constructed by merging all linked BNs and linkages into one graph  DAG  Variable nodes, Linkage Nodes  Edges: all edges in the linked BNs have a representation in J-graph  CPT: Q(A|  A ) = P(A|  A ), Q(A|B) = P S (A|B) for ► Q: distribution in J-graph, P: original distribution

J-Graph (3) Example A BC D A’ B’ C’ D’ A1A1 B→B’; 1→2 C→C’; 1→2 D2D2 A’ 2 D’ 2  Linkage nodes  represent all linked variables and the linkage  encode the similarity of the linkage in CPT  merge the CPTs by IPFP Linkage Node

Concept Mapping using SLBN (1) Motivations ► Ontology mappings are seldom certain  Existing approaches ► use hard threshold to filter mappings ► throw similarities away after mappings are created ► mappings are identical and 1-1  But ► often one concept is similar to more than one concept ► Semantically similar concepts are hard to be represented logically

Probabilistic Information Learner BayesOWL Concept Mapping using SLBN (2) The Framework Onto 2 Onto 1 WWW BN 2 BN 1 Variable Linkages BayesOWL SLBN

Concept Mapping using SLBN (3) Objection ► Discover new and complex concept mappings  Make full use of the learned similarity in SLBN’s inference  Create an expression for a concept in another ontology ► Find how similar “Onto1:B  Onto1:C” is to “Onto2:A” ► Experiments have shown encouraging results

Concept Mapping using SLBN (3) Experiment ► Artificial Intelligence sub-domain from ACM Topic Taxonomy DMOZ (Open Directory) hierarchies Learned Similarities: J(dmoz.sw, acm.rs) = 0.64 J(dmoz.sw, acm.sn) = 0.61 J(dmoz.sw, acm.krfm) = 0.49 After SLBN Inference: J(dmoz.sw, acm.rs  acm.sn) = Q (acm.rs = True  acm.sn = True | dmoz.sw = True) =

Future Works ► Modeling with SLBN  Discover semantic similar concepts by machine learning algorithms  Create effective and correct linkages from learned algorithms ► Distributed Inference methods ► Loosing the restrictions  Inference with linkages of both directions  Use functions to represent similarities

Thank You! ► Questions?

Background Semantics of BN ► Chain rule where  (a i ) is the parent set of a i. ► d-separation: B CA CA B A B C serials diverging converging Instantiated Not instantiated d-separated variables do not influence each other.