I-equivalence Bayesian Networks Representation Probabilistic Graphical

Slides:



Advertisements
Similar presentations
Local structures; Causal Independence, Context-sepcific independance COMPSCI 276 Fall 2007.
Advertisements

Bayesian Networks. Contents Semantics and factorization Reasoning Patterns Flow of Probabilistic Influence.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Bayesian Networks A causal probabilistic network, or Bayesian network,
Parameter Learning in Markov Nets Dhruv Batra, Recitation 11/13/2008.
1Causality & MDL Causal Models as Minimal Descriptions of Multivariate Systems Jan Lemeire June 15 th 2006.
Artificial Intelligence and Lisp Lecture 7 LiU Course TDDC65 Autumn Semester, 2010
Back-Propagation Algorithm
Associative Learning.
December 7, 2010Neural Networks Lecture 21: Hopfield Network Convergence 1 The Hopfield Network The nodes of a Hopfield network can be updated synchronously.
Approximate Inference 2: Monte Carlo Markov Chain
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
机器学习 陈昱 北京大学计算机科学技术研究所 信息安全工程研究中心. Concept Learning Reference : Ch2 in Mitchell’s book 1. Concepts: Inductive learning hypothesis General-to-specific.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Sum-Product Networks CS886 Topics in Natural Language Processing
Dan Boneh Symmetric Encryption History Crypto. Dan Boneh History David Kahn, “The code breakers” (1996)
Lectures 2 – Oct 3, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall.
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
Bayesian Network By Zhang Liliang. Key Point Today Intro to Bayesian Network Usage of Bayesian Network Reasoning BN: D-separation.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2008 Readings: K&F: 3.1, 3.2, –  Carlos.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Daphne Koller Wrapup BNs vs MNs Probabilistic Graphical Models Representation.
Additional NN Models Reinforcement learning (RL) Basic ideas: –Supervised learning: (delta rule, BP) Samples (x, f(x)) to learn function f(.) precise error.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Daphne Koller Template Models Plate Models Probabilistic Graphical Models Representation.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Reasoning Patterns Bayesian Networks Representation Probabilistic
Integers on a Number Line Grade 6 Standard 1.C.1.a.
Daphne Koller Introduction Motivation and Overview Probabilistic Graphical Models.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Daphne Koller Independencies Bayesian Networks Probabilistic Graphical Models Representation.
Maximum Expected Utility
Ch7: Hopfield Neural Model
Context-Specific CPDs
A Simple Artificial Neuron
Neural Networks for Vertex Covering
General Gibbs Distribution
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
Preliminaries: Distributions
Luger: Artificial Intelligence, 5th edition
Bayesian Networks Independencies Representation Probabilistic
Independence in Markov Networks
General Gibbs Distribution
CS 621 Artificial Intelligence Lecture 25 – 14/10/05
Convolutional networks
Readings: K&F: 5.1, 5.2, 5.3, 5.4, 5.5, 5.6, 5.7 Markov networks, Factor graphs, and an unified view Start approximate inference If we are lucky… Graphical.
MCMC for PGMs: The Gibbs Chain
Conditional Random Fields
Probabilistic Influence & d-separation
Reasoning Patterns Bayesian Networks Representation Probabilistic
Factorization & Independence
Factorization & Independence
Unifying Variational and GBP Learning Parameters of MNs EM for BNs
Conditional Random Fields
Markov Networks Independencies Representation Probabilistic Graphical
Tree-structured CPDs Local Structure Representation Probabilistic
Crypto Encryption Intro to public key.
Representation Probabilistic Graphical Models Local Structure Overview.
Independence in Markov Networks
Markov Networks Independencies Representation Probabilistic Graphical
Probabilistic Reasoning
Flow of Probabilistic Influence
Preliminaries: Independence
Use the Chain Rule to find {image} where t = 5, u = 3, v = 0. {image}
Presentation transcript:

I-equivalence Bayesian Networks Representation Probabilistic Graphical Models Bayesian Networks I-equivalence

Different G’s might encode the same independencies Draw student network

Which of the following graphs does not encode the same independencies as the others? X Y Z X Y Z X Y Z X Y Z

I-equivalence Definition: Two graphs G1 and G2 over X1,…,Xn are I-equivalent if Chain rule, CPDs

Characterizing I-equivalence Theorem: G1 and G2 are I-equivalent if and only if they have

Implications

END END END

The Chain Rule for Bayesian Nets Intelligence Difficulty Grade Letter SAT 0.3 0.08 0.25 0.4 g2 0.02 0.9 i1,d0 0.7 0.05 i0,d1 0.5 g1 g3 0.2 i1,d1 i0,d0 l1 l0 0.99 0.1 0.01 0.6 0.95 s0 s1 0.8 i1 i0 d1 d0 P(D,I,G,S,L) = P(D) P(I) P(G | I,D) P(L | G) P(S | I)

Suppose q is at a local minimum of a function Suppose q is at a local minimum of a function. What will one iteration of gradient descent do? Leave q unchanged. Change q in a random direction. Move q towards the global minimum of J(q). Decrease q.

Consider the weight update: Which of these is a correct vectorized implementation?

Fig. A corresponds to a=0.01, Fig. B to a=0.1, Fig. C to a=1.