Topics Combining probability and first-order logic –BLOG and DBLOG Learning very complex behaviors –ALisp: hierarchical RL with partial programs State.

Slides:



Advertisements
Similar presentations
Online Filtering, Smoothing & Probabilistic Modeling of Streaming Data In short, Applying probabilistic models to Streams Bhargav Kanagal & Amol Deshpande.
Advertisements

Gibbs sampling in open-universe stochastic languages Nimar S. Arora Rodrigo de Salvo Braz Erik Sudderth Stuart Russell.
Rao-Blackwellised Particle Filtering Based on Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks by Arnaud Doucet, Nando de Freitas, Kevin.
Introduction of Probabilistic Reasoning and Bayesian Networks
Artificial Intelligence Lecture 2 Dr. Bo Yuan, Professor Department of Computer Science and Engineering Shanghai Jiaotong University
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
Plan Recognition with Multi- Entity Bayesian Networks Kathryn Blackmond Laskey Department of Systems Engineering and Operations Research George Mason University.
WebMiningResearch ASurvey Web Mining Research: A Survey Raymond Kosala and Hendrik Blockeel ACM SIGKDD, July 2000 Presented by Shan Huang, 4/24/2007.
CSE 574 – Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
CSE 574: Artificial Intelligence II Statistical Relational Learning Instructor: Pedro Domingos.
1 Department of Computer Science and Engineering, University of South Carolina Issues for Discussion and Work Jan 2007  Choose meeting time.
Dynamic Bayesian Networks CSE 473. © Daniel S. Weld Topics Agency Problem Spaces Search Knowledge Representation Reinforcement Learning InferencePlanningLearning.
Human-level AI Stuart Russell Computer Science Division UC Berkeley.
CSE 515 Statistical Methods in Computer Science Instructor: Pedro Domingos.
Conference Paper by: Bikramjit Banerjee University of Southern Mississippi From the Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence.
Research paper: Web Mining Research: A survey SIGKDD Explorations, June Volume 2, Issue 1 Author: R. Kosala and H. Blockeel.
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Relational Probability Models Brian Milch MIT 9.66 November 27, 2007.
First-Order Probabilistic Languages: Into the Unknown Brian Milch and Stuart Russell University of California at Berkeley, USA August 27, 2006 Based on.
From Bayesian Filtering to Particle Filters Dieter Fox University of Washington Joint work with W. Burgard, F. Dellaert, C. Kwok, S. Thrun.
IJCAI 2003 Workshop on Learning Statistical Models from Relational Data First-Order Probabilistic Models for Information Extraction Advisor: Hsin-His Chen.
K. J. O’Hara AMRS: Behavior Recognition and Opponent Modeling Oct Behavior Recognition and Opponent Modeling in Autonomous Multi-Robot Systems.
Representational and inferential foundations for possible large-scale information extraction and question-answering from the web Stuart Russell Computer.
Markov Logic And other SRL Approaches
Unknown Objects and BLOG Brian Milch MIT IPAM Summer School July 16, 2007.
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
Announcements  Project 3 will be posted ASAP: hunting ghosts with sonar.
Artificial Intelligence: Introduction Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.
Decision Making Under Uncertainty Lec #8: Reinforcement Learning UIUC CS 598: Section EA Professor: Eyal Amir Spring Semester 2006 Most slides by Jeremy.
1 2010/2011 Semester 2 Introduction: Chapter 1 ARTIFICIAL INTELLIGENCE.
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
4. Particle Filtering For DBLOG PF, regular BLOG inference in each particle Open-Universe State Estimation with DBLOG Rodrigo de Salvo Braz*, Erik Sudderth,
BLOG: Probabilistic Models with Unknown Objects Brian Milch, Bhaskara Marthi, Stuart Russell, David Sontag, Daniel L. Ong, Andrey Kolobov University of.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Latent Dirichlet Allocation
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Representational and inferential foundations for possible large-scale information extraction and question-answering from the web Stuart Russell Computer.
First-Order Probabilistic Inference Rodrigo de Salvo Braz.
Probability and Time. Overview  Modelling Evolving Worlds with Dynamic Baysian Networks  Simplifying Assumptions Stationary Processes, Markov Assumption.
1 Chapter 17 2 nd Part Making Complex Decisions --- Decision-theoretic Agent Design Xin Lu 11/04/2002.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
BLOG: Probabilistic Models with Unknown Objects Brian Milch Harvard CS 282 November 29,
Learning and Structural Uncertainty in Relational Probability Models Brian Milch MIT 9.66 November 29, 2007.
GENETIC ALGORITHMS Tanmay, Abhijit, Ameya, Saurabh.
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
First-Order Probabilistic Inference Rodrigo de Salvo Braz SRI International.
The Approach of Modern AI Stuart Russell Computer Science Division UC Berkeley.
Modeling of Core Protection Calculator System Software February 28, 2005 Kim, Sung Ho Kim, Sung Ho.
Probabilistic Reasoning Inference and Relational Bayesian Networks.
CS 541: Artificial Intelligence Lecture VIII: Temporal Probability Models.
Announcements  Upcoming due dates  Thursday 10/1 in class Midterm  Coverage: everything in lecture and readings except first-order logic; NOT probability.
Chapter 9: Means and Proportions as Random Variables 9.1 Understanding dissimilarity among samples 9.2 Sampling distributions for sample proportions 9.3.
NTT-MIT Collaboration Meeting, 2001Leslie Pack Kaelbling 1 Learning in Worlds with Objects Leslie Pack Kaelbling MIT Artificial Intelligence Laboratory.
Introduction to Sampling based inference and MCMC
Machine Learning Applications in Grid Computing
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Uncertainty in an Unknown World
Artificial Intelligence introduction(2)
CSE 515 Statistical Methods in Computer Science
Open universes and nuclear weapons
Machine learning, probabilistic modelling
Computational Physiology for Critical Care Monitoring
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11
Algorithms and Problem Solving
CS 188: Artificial Intelligence
Representations & Reasoning Systems (RRS) (2.2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
The open universe Stuart Russell
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

Topics Combining probability and first-order logic –BLOG and DBLOG Learning very complex behaviors –ALisp: hierarchical RL with partial programs State estimation for neurotrauma patients –Joint w/ Geoff Manley (UCSF), Intel, Omron Metareasoning and bounded optimality Transfer learning (+ Jordan, Bartlett, MIT, SU, OSU) Knowing everything on the web Human-level AI

Heart rate Blood pressure Oxygen saturation Pulmonary artery pressure Intracranial pressure Temperature Nursing Documentation MedicationsTreatments Intake Output Vital signs Blood products IV fluids Tissue oxygen Inspired oxygen Tidal volume Peak pressure End expiratory pressure Respiratory rate Ventilation mode Cardiac output Sedation level ICP wave State estimation: 3x5 index card

Patient 2-13

Dynamic Bayesian Networks

DBNs contd:

Research plan DBN model: ~200 core state variables, ~500 sensor-related variables Learn model parameter distributions from DB Infer patient-specific parameters online Goals: –Improved alarms –Diagnostic state estimation => improved treatment –Solve the treatment POMDP –Structure discovery => better understanding of physiology

Possible worlds Propositional First-order + unique names, domain closure First-order open-world AB CD AB CD AB CD AB CD AB CD A B C D

Example: Citation Matching [Lashkari et al 94] Collaborative Interface Agents, Yezdi Lashkari, Max Metral, and Pattie Maes, Proceedings of the Twelfth National Conference on Articial Intelligence, MIT Press, Cambridge, MA, Metral M. Lashkari, Y. and P. Maes. Collaborative interface agents. In Conference of the American Association for Artificial Intelligence, Seattle, WA, August Are these descriptions of the same object? What authors and papers actually exist, with what attributes? Who wrote which papers? General problem: raw data -> relational KB Other examples: multitarget tracking, vision, NLP Approach: formal language for specifying first-order open-world probability models

BLOG generative process Number statements describe steps that add some objects to the world Dependency statements describe steps that set the value of a function or relation on a tuple of arguments Includes setting the referent of a constant symbol (0-ary function) Both types may condition on existence and properties of previously added objects

BLOG model (simplified) guaranteed Citation Cit1, Cit2, Cit3, Cit4, Cit5, Cit6, Cit7; #Researcher ~ NumResearchersPrior(); Name(r) ~ NamePrior(); #Paper(FirstAuthor = r) ~ NumPapersPrior(Position(r)); Title(p) ~ TitlePrior(); PubCited(c) ~ Uniform({Paper p}); Text(c) ~ NoisyCitationGrammar (Name(FirstAuthor(PubCited(c))), Title(PubCited(c)));

Basic results Theorem 1: Every well-formed* BLOG model specifies a unique distribution over possible worlds The probability of each (finite) world is given by a product of the relevant conditional probabilities from the model Theorem 2: For any well-formed BLOG model, there are algorithms (LW, MCMC) that converge to correct probability for any query, using finite time per sampling step

Citation Matching Results Four data sets of ~ citations, referring to ~ papers

DBLOG BLOG allows for temporal models – time is just a logical variable over an infinite set Inference works (only finitely many relevant random variables) but is grossly inefficient DBLOG includes time as a distinguished type and predecessor as distinguished function; implements special-purpose inference: –Particle filter for temporally varying relations –Decayed MCMC for atemporal relations

Open Problems Inference –Applying “lifted” inference to BLOG (like Prolog) –Approximation algorithms for problems with huge/growing numbers of objects Knowledge representation –Hierarchical activity models –Undirected submodels –Nonparametric extensions (cf. de Freitas, 2005)