Reasoning Patterns Bayesian Networks Representation Probabilistic

Slides:



Advertisements
Similar presentations
Deep Learning Bing-Chen Tsai 1/21.
Advertisements

Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
Local structures; Causal Independence, Context-sepcific independance COMPSCI 276 Fall 2007.
Bayesian Networks. Contents Semantics and factorization Reasoning Patterns Flow of Probabilistic Influence.
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Probabilistic Reasoning with Uncertain Data Yun Peng and Zhongli Ding, Rong Pan, Shenyong Zhang.
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
1 © 1998 HRL Laboratories, LLC. All Rights Reserved Development of Bayesian Diagnostic Models Using Troubleshooting Flow Diagrams K. Wojtek Przytula: HRL.
Approximate Inference 2: Monte Carlo Markov Chain
Biointelligence Laboratory, Seoul National University
1 Approximate Inference 2: Importance Sampling. (Unnormalized) Importance Sampling.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
Mapping and Localization with RFID Technology Matthai Philipose, Kenneth P Fishkin, Dieter Fox, Dirk Hahnel, Wolfram Burgard Presenter: Aniket Shah.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Dan Boneh Symmetric Encryption History Crypto. Dan Boneh History David Kahn, “The code breakers” (1996)
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
Made by: Maor Levy, Temple University  Inference in Bayes Nets ◦ What is the probability of getting a strong letter? ◦ We want to compute the.
Bayesian Network By Zhang Liliang. Key Point Today Intro to Bayesian Network Usage of Bayesian Network Reasoning BN: D-separation.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
CHAPTER 4, Part II Oliver Schulte Summer 2011 Local Search.
CPSC 422, Lecture 11Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 2, 2015.
Wei Sun and KC Chang George Mason University March 2008 Convergence Study of Message Passing In Arbitrary Continuous Bayesian.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
Additional NN Models Reinforcement learning (RL) Basic ideas: –Supervised learning: (delta rule, BP) Samples (x, f(x)) to learn function f(.) precise error.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
CHAPTER 3: BAYESIAN DECISION THEORY. Making Decision Under Uncertainty Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Daphne Koller Template Models Plate Models Probabilistic Graphical Models Representation.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Daphne Koller Bayesian Networks Semantics & Factorization Probabilistic Graphical Models Representation.
Reasoning Patterns Bayesian Networks Representation Probabilistic
Daphne Koller Introduction Motivation and Overview Probabilistic Graphical Models.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Daphne Koller Independencies Bayesian Networks Probabilistic Graphical Models Representation.
MLPR - Questions. Can you go through integration, differentiation etc. Why do we need priors? Difference between prior and posterior. What does Bayesian.
Maximum Expected Utility
Ch7: Hopfield Neural Model
CSC321: Neural Networks Lecture 19: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Context-Specific CPDs
Bayesian Biosurveillance of Disease Outbreaks
Probabilistic Models for Linear Regression
Bayesian Network Reasoning with Gibbs Sampling
CSE-490DF Robotics Capstone
Neural Networks for Vertex Covering
Learning Markov Networks
General Gibbs Distribution
of the Artificial Neural Networks.
Preliminaries: Distributions
Luger: Artificial Intelligence, 5th edition
Bayesian Statistics and Belief Networks
Introduction to Neural Networks
Graduate School of Information Sciences, Tohoku University
Simple Sampling Sampling Methods Inference Probabilistic Graphical
I-equivalence Bayesian Networks Representation Probabilistic Graphical
Shared Features in Log-Linear Models
MCMC for PGMs: The Gibbs Chain
Conditional Random Fields
Probabilistic Influence & d-separation
Factorization & Independence
Factorization & Independence
Crypto Encryption Intro to public key.
Plate Models Template Models Representation Probabilistic Graphical
Probabilistic Reasoning
Plate Models Template Models Representation Probabilistic Graphical
Flow of Probabilistic Influence
Preliminaries: Independence
Variable Elimination Graphical Models – Carlos Guestrin
Presentation transcript:

Reasoning Patterns Bayesian Networks Representation Probabilistic Graphical Models Bayesian Networks Reasoning Patterns

The Student Network 0.4 0.6 d1 d0 0.3 0.7 i1 i0 Difficulty Intelligence 0.2 0.95 s0 s1 0.8 i1 0.05 i0 0.3 0.08 0.25 0.4 g2 0.02 0.9 i1,d0 0.7 0.05 i0,d1 0.5 g1 g3 0.2 i1,d1 i0,d0 Grade SAT Letter l1 l0 0.99 0.4 0.1 0.9 g1 0.01 g3 0.6 g2

Causal Reasoning P(l1) ~ 0.5 P(l1 | i0 ) ~ P(l1 | i0 , d0) ~ Difficulty Difficulty Intelligence Intelligence Grade SAT P(l1) ~ 0.5 Letter P(l1 | i0 ) ~ P(l1 | i0 , d0) ~

Evidential Reasoning P(d1) = 0.4 P(i1) = 0.3 P(d1 | g3) ≈ P(i1 | g3) ≈ Difficulty Intelligence Student gets a C  Grade SAT 0.3 0.08 0.25 0.4 g2 0.02 0.9 i1,d0 0.7 0.05 i0,d1 0.5 g1 g3 0.2 i1,d1 i0,d0 Letter 0.63, 0.08

We find out that class is hard What happens to the posterior probability of high intelligence? Intelligence Difficulty Grade Letter SAT Class is hard! Student gets a C  Goes up Goes down Doesn’t change We can’t know

Intercausal Reasoning P(d1) = 0.4 P(i1) = 0.3 P(d1 | g3) ≈ 0.63 P(i1 | g3) ≈ 0.08 P(i1 | g3, d1) ≈ 0.11 Difficulty Intelligence Class is hard! Grade SAT Student gets a C  Letter 0.11

Intercausal Reasoning II P(i1) = 0.3 P(i1 | g2) ≈ P(i1 | g2, d1) ≈ Difficulty Difficulty Intelligence Class is hard! Grade SAT Student gets a B  Letter 0.175, 0.34

Student Aces the SAT What happens to the posterior probability that the class is hard? Intelligence Difficulty Grade Letter SAT Student gets a C  Goes up Goes down Doesn’t change Student aces the SAT  We can’t know

Multiple Evidence P(d1) = 0.4 P(i1) = 0.3 P(d1 | g3) ≈ 0.63 P(i1 | g3) ≈ 0.08 P(d1 | g3, s1) ≈ P(i1 | g3, s1) ≈ Difficulty Intelligence Grade SAT Student gets a C  Student aces the SAT  Letter 0.76, 0.58

END

Suppose q is at a local minimum of a function Suppose q is at a local minimum of a function. What will one iteration of gradient descent do? Leave q unchanged. Change q in a random direction. Move q towards the global minimum of J(q). Decrease q.

Consider the weight update: Which of these is a correct vectorized implementation?

Fig. A corresponds to a=0.01, Fig. B to a=0.1, Fig. C to a=1.