ICS 280 Learning in Graphical Models

Slides:



Advertisements
Similar presentations
Bayes rule, priors and maximum a posteriori
Advertisements

INTRODUCTION TO MACHINE LEARNING Bayesian Estimation.
Pattern Recognition and Machine Learning
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
Supervised Learning Recap
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Bayesian Wrap-Up (probably). 5 minutes of math... Marginal probabilities If you have a joint PDF:... and want to know about the probability of just one.
1 Graphical Models in Data Assimilation Problems Alexander Ihler UC Irvine Collaborators: Sergey Kirshner Andrew Robertson Padhraic Smyth.
An Introduction to Bayesian Inference Michael Betancourt April 8,
Machine Learning CMPT 726 Simon Fraser University CHAPTER 1: INTRODUCTION.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Presenting: Assaf Tzabari
Parametric Inference.
Machine Learning CMPT 726 Simon Fraser University
Rutgers CS440, Fall 2003 Introduction to Statistical Learning Reading: Ch. 20, Sec. 1-4, AIMA 2 nd Ed.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Learning In Bayesian Networks. General Learning Problem Set of random variables X = {X 1, X 2, X 3, X 4, …} Training set D = { X (1), X (2), …, X (N)
BCS547 Neural Decoding.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
IE 300, Fall 2012 Richard Sowers IESE. 8/30/2012 Goals: Rules of Probability Counting Equally likely Some examples.
Lecture 2: Statistical learning primer for biologists
Dropout as a Bayesian Approximation
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Machine Learning Basics ( 1/2 ) 周岚. Machine Learning Basics What do we mean by learning? Mitchell (1997) : A computer program is said to learn from experience.
Stat 223 Introduction to the Theory of Statistics
Statistical Estimation
Chapter 3: Maximum-Likelihood Parameter Estimation
DEEP LEARNING BOOK CHAPTER to CHAPTER 6
HST 583 fMRI DATA ANALYSIS AND ACQUISITION
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Ch3: Model Building through Regression
Parameter Estimation 主講人:虞台文.
Special Topics In Scientific Computing
Distributions and Concepts in Probability Theory
CSCI 5822 Probabilistic Models of Human and Machine Learning
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
Statistical NLP: Lecture 4
CSCI 5822 Probabilistic Models of Human and Machine Learning
Pattern Recognition and Machine Learning
Stat 223 Introduction to the Theory of Statistics
LECTURE 07: BAYESIAN ESTIMATION
Parametric Methods Berlin Chen, 2005 References:
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions July chonbuk national university.
CS639: Data Management for Data Science
12. Principles of Parameter Estimation
Mathematical Foundations of BME Reza Shadmehr
Maximum Likelihood Estimation (MLE)
Presentation transcript:

ICS 280 Learning in Graphical Models Time: Tu-Th, 5-6.20 pm Instructor: Max Welling

Further Information URL: http://www.ics.uci.edu/~welling/teaching/GraphicalModels.html check for readings, homework and other updates. Prerequisites: ICS 274 Probabilistic Learning: Theory and Algorithms, or with consent of instructor.

Overview review of statistical concepts introduction to graphical models hidden variables and exact inference learning in graphical models unsupervised learning supervised learning graphical models of time series approximate inference Bayesian learning and structure learning

Review of Statistical Concepts

Basic Definitions definition probability distribution, discrete/continuous joint, conditional & marginal distributions independence Bayes rule moments: mean & covariance multivariate Gaussian distribution learning, supervised & unsupervised observed and unobserved random variables

Bayesian Estimation parameters are random variables “learning” = computing the posterior distribution of the parameters given the data. priors and hyper-parameters the marginal likelihood or evidence the predictive distribution example & demo_Bayes

MAP Estimation MAP = maximum a posteriori value of the parameter much easier to compute throw away uncertainty in estimate of the parameter prior distribution still has an impact example & demo_MAP

Maximum Likelihood MAP with uniform prior (misleading) frequentist versus Baysian ML is an estimator which has “nice” properties. Bias-variance trade-off ML is asymptotic unbiased and efficient (minimal variance) Cramer-Rao bound Gaussian example & demo_ML

Generalization & Overfitting What is overfitting ? ML is not protected against overfitting MAP is partially protected against overfitting Baysian estimation is completely protected against overfitting Adding regularization terms to ML Minimum description length: the coding perspective