Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
A Tutorial on Learning with Bayesian Networks
Review of Probability. Definitions (1) Quiz 1.Let’s say I have a random variable X for a coin, with event space {H, T}. If the probability P(X=H) is.
Bayesian Estimation in MARK
Dynamic Bayesian Networks (DBNs)
Part 24: Bayesian Estimation 24-1/35 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Introduction of Probabilistic Reasoning and Bayesian Networks
CHAPTER 16 MARKOV CHAIN MONTE CARLO
BAYESIAN INFERENCE Sampling techniques
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Exact Inference (Last Class) variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Bayesian Networks. Graphical Models Bayesian networks Conditional random fields etc.
Artificial Intelligence and Lisp Lecture 7 LiU Course TDDC65 Autumn Semester, 2010
End of Chapter 8 Neil Weisenfeld March 28, 2005.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
CS 188: Artificial Intelligence Fall 2006 Lecture 17: Bayes Nets III 10/26/2006 Dan Klein – UC Berkeley.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Announcements Homework 8 is out Final Contest (Optional)
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Lecture II-2: Probability Review
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Introduction to Bayesian Networks
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Randomized Algorithms for Bayesian Hierarchical Clustering
Bayes’ Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Exact Inference (Last Class) Variable elimination  polytrees (directed graph with at most one undirected path between any two vertices; subset of DAGs)
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
Lecture 2: Statistical learning primer for biologists
Probabilistic models Jouni Tuomisto THL. Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
CS 188: Artificial Intelligence Bayes Nets: Approximate Inference Instructor: Stuart Russell--- University of California, Berkeley.
Exact Inference in Bayes Nets. Notation U: set of nodes in a graph X i : random variable associated with node i π i : parents of node i Joint probability:
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Pattern Recognition and Machine Learning
CPSC 7373: Artificial Intelligence Lecture 5: Probabilistic Inference Jiang Bian, Fall 2012 University of Arkansas at Little Rock.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
CS 416 Artificial Intelligence Lecture 15 Uncertainty Chapter 14 Lecture 15 Uncertainty Chapter 14.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
CS 541: Artificial Intelligence Lecture VII: Inference in Bayesian Networks.
INTRODUCTION TO Machine Learning 2nd Edition
CS 2750: Machine Learning Directed Graphical Models
Advanced Statistical Computing Fall 2016
CS 4/527: Artificial Intelligence
Prof. Adriana Kovashka University of Pittsburgh April 4, 2017
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
CAP 5636 – Advanced Artificial Intelligence
Markov Networks.
Predictive distributions
Instructors: Fei Fang (This Lecture) and Dave Touretzky
Ch13 Empirical Methods.
CS 188: Artificial Intelligence
Class #19 – Tuesday, November 3
Econometrics Chengyuan Yin School of Mathematics.
CS 188: Artificial Intelligence Fall 2008
Class #16 – Tuesday, October 26
CS 188: Artificial Intelligence Fall 2007
Markov Networks.
Presentation transcript:

Probabilistic models Jouni Tuomisto THL

Outline Deterministic models with probabilistic parameters Hierarchical Bayesian models Bayesian belief nets

Deterministic models with probabilistic parameters Inputs are uncertain, but causal relations are assumed certain. Works well with established situations, especially if physical foundations. Exposure = ∑ (c i t i ) / ∑ t i –i = microenvironment –c = concentration –t = time

Functional vs. probabilistic dependency Va1=2.54*Ch1^2 Va2=normal(2.54*Ch1^2,2)

Hierarchical Bayesian models Relations are probabilistic Gibbs Sampler Another MCMC (Markov chain Monte Carlo) Method Update a single parameter at a time Sample from conditional distribution When other parameters are fixed

Gibbs sampling To introduce the Gibbs sampler, consider a bivariate random variable (x; y), and suppose we wish to compute one or both marginals, p(x) and p(y). The idea behind the sampler is that it is far easier to consider a sequence of conditional distributions, p(x | y) and p(y | x), than it is to obtain the marginal by integration of the joint density p(x; y), e.g., –p(x) = ∫ p(x; y)dy.

Gibbs sampling in practice The sampler starts with some initial value y0 for y and obtains x0 by generating a random variable from the conditional distribution p(x | y = y0). The sampler then uses x0 to generate a new value of y1, drawing from the conditional distribution based on the value x0, p(y j x = x0). The sampler proceeds as follows xi ≈ p(x | y = yi-1) (proportionality) yi ≈ p(y | x = xi) Repeating this process k times, generates a Gibbs sequence of length k, where a subset of points (xj; yj) for 1 ≤ j ≤ m < k are taken as our simulated draws from the full joint distribution.

Hierarchical model with parameters and hyperparameters A useful graphical tool for representing hierarchical Bayes models is the directed acyclic graph, or DAG. In this diagram, the likelihood function is represented as the root of the graph; each prior is represented as a separate node pointing to the node that depends on it.directed acyclic graph

Bayesian belief nets Relations are described either with conditional probabilities P(x|y), P(y) or marginal probabilities P(x), P(y) and a rank correlation between them. You need to get the conditional probabilities from somewhere. –Unlike hierarchical Bayes model, belief nets are not developed for updating when new data comes out. The model is used to make inference.

Bayesian belief nets P(sprinkler | rain) P(grass wet | sprinkler, rain) P(rain)

Uninet: diagram view V1: rain (mm/day) V2: sprinkler on (h/day) V3 ”wetness” of grass (range 0-1)

Uninet: variable definition view

Bayes belief network: unconditional situation

Conditioning on input variables

Conditioning on outcome