Probabilistic Models in Human and Machine Intelligence

Slides:



Advertisements
Similar presentations
Approaches, Tools, and Applications Islam A. El-Shaarawy Shoubra Faculty of Eng.
Advertisements

PAC Learning 8/5/2005. purpose Effort to understand negative selection algorithm from totally different aspects –Statistics –Machine learning What is.
The influence of domain priors on intervention strategy Neil Bramley.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
The Logic of Intelligence Pei Wang Department of Computer and Information Sciences Temple University.
Rulebase Expert System and Uncertainty. Rule-based ES Rules as a knowledge representation technique Type of rules :- relation, recommendation, directive,
Probabilistic Models in Human and Machine Intelligence.
Computation and representation Joe Lau. Overview of lecture What is computation? Brief history Computational explanations in cognitive science Levels.
Neural Network Intro Slides
AI 授課教師:顏士淨 2013/09/12 1. Part I & Part II 2  Part I Artificial Intelligence 1 Introduction 2 Intelligent Agents Part II Problem Solving 3 Solving Problems.
Bart Selman CS CS 475: Uncertainty and Multi-Agent Systems Prof. Bart Selman Introduction.
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
Probabilistic Models of Cognition Conceptual Foundations Chater, Tenenbaum, & Yuille TICS, 10(7), (2006)
Probabilistic inference in human semantic memory Mark Steyvers, Tomas L. Griffiths, and Simon Dennis 소프트컴퓨팅연구실오근현 TRENDS in Cognitive Sciences vol. 10,
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
CS 547: Sensing and Planning in Robotics Gaurav S. Sukhatme Computer Science Robotic Embedded Systems Laboratory University of Southern California
CS 357 – Intro to Artificial Intelligence  Learn about AI, search techniques, planning, optimization of choice, logic, Bayesian probability theory, learning,
Exploring subjective probability distributions using Bayesian statistics Tom Griffiths Department of Psychology Cognitive Science Program University of.
Tom Griffiths CogSci C131/Psych C123 Computational Models of Cognition.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Priors and predictions in everyday cognition Tom Griffiths Cognitive and Linguistic Sciences.
CSE 471/598,CBS598 Introduction to Artificial Intelligence Fall 2004
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
LLNL-PRES This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
CISC4/681 Introduction to Artificial Intelligence1 Introduction – Artificial Intelligence a Modern Approach Russell and Norvig: 1.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Optimal predictions in everyday cognition Tom Griffiths Josh Tenenbaum Brown University MIT Predicting the future Optimality and Bayesian inference Results.
A Brief Introduction to GA Theory. Principles of adaptation in complex systems John Holland proposed a general principle for adaptation in complex systems:
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Transfer Learning Task. Problem Identification Dataset : A Year: 2000 Features: 48 Training Model ‘M’ Testing 98.6% Training Model ‘M’ Testing 97% Dataset.
Perceptual Multistability as Markov Chain Monte Carlo Inference.
Randomized Algorithms for Bayesian Hierarchical Clustering
Review: Probability Random variables, events Axioms of probability Atomic events Joint and marginal probability distributions Conditional probability distributions.
Probabilistic Models in Human and Machine Intelligence.
International Conference on Fuzzy Systems and Knowledge Discovery, p.p ,July 2011.
Gaussian Processes For Regression, Classification, and Prediction.
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
A Brief History of AI Fall 2013 COMP3710 Artificial Intelligence Computing Science Thompson Rivers University.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
From NARS to a Thinking Machine Pei Wang Temple University.
1 Artificial Intelligence & Prolog Programming CSL 302.
Pattern Recognition. What is Pattern Recognition? Pattern recognition is a sub-topic of machine learning. PR is the science that concerns the description.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Bayesian inference in neural networks
CS440/ECE448: Artificial Intelligence Lecture 1: What is AI?
A probabilistic approach to cognition
CHAPTER 5 Handling Uncertainty BIC 3337 EXPERT SYSTEM.
What is cognitive psychology?
School of Computer Science & Engineering
Done Done Course Overview What is AI? What are the Major Challenges?
Markov ó Kalman Filter Localization
Course: Autonomous Machine Learning
CSCI 5822 Probabilistic Models of Human and Machine Learning
(Very Brief) Introduction to Bayesian Statistics
CSE-490DF Robotics Capstone
CSCI 5822 Probabilistic Models of Human and Machine Learning
Probability and Time: Markov Models
EA C461 – Artificial Intelligence Introduction
Prepared by: Mahmoud Rafeek Al-Farra
Statistical Data Mining
Probability and Time: Markov Models
COMP3710 Artificial Intelligence Thompson Rivers University
Toward a Great Class Project: Discussion of Stoianov & Zorzi’s Numerosity Model Psych 209 – 2019 Feb 14, 2019.
Presentation transcript:

Probabilistic Models in Human and Machine Intelligence CSCI 5822 Spring 2018

Current Class Status 67 enrolled 12 waitlisted

Machine Learning @ CU Intro courses Other courses this semester CSCI 5622: Machine Learning CSCI 5352: Network Analysis and Modeling CSCI 5822: Probabilistic Models CSCI 5922: Neural Networks and Deep Learning Other courses this semester CSCI 7000-010: Mind Reading Machines (Sidney D’Mello) CSCI 7000-008: Human Centered ML (Chenhao Tan)

A Very Brief History of Cog Sci and AI 1950’s-1980’s Symbolic computation The mind is like a modern digital computer Symbols are basic elements of representation, e.g., John, father Symbol manipulation is basic operation, e.g., (father Y X) & (father Z X) -> (sibling Y Z) 1980’s-1990’s Subsymbolic computation The mind is a massively parallel network of simple neuron-like processors Numerical vectors are basic elements of representation Numerical computing is basic operation: y = f ( Σi wi xi) 2000’s – 2010’s Probabilistic computation The mind operates according to laws of probabilistic reasoning and statistical inference Invades cog sci, AI (planning, natural language processing), ML Formalizes the statistical intuitions underlying neural nets 2010’s and beyond Subsymbolic computation The mind is a massively parallel neuron-like network of simple processors… Well grounded in probability and statistics (e.g., Variational autoencoders, Boltzmann machines) ML conference in 1986 – all symbolic Explanation based learning was the rage

Relation of Probabilistic Models to Symbolic and Subsymbolic Models statistical learning (large # examples) rule learning (small # examples) feature-vector representations structured representations structured rep: father of your brother is your father

What Is Probability? Frequentist notion Relative frequency obtained if event were observed many times (e.g., coin flip) Subjective notion Degree of belief in some hypothesis Analogous to neural net activation Long philosophical battle between these two views Subjective notion makes sense for cog sci and AI given that probabilities represent mental states

Do People Reason According To The Laws Of Probability? The probability of breast cancer is 1% for a woman at 40 who participates in routine screening. If a woman has breast cancer, the probability is 80% that she will have a positive mammography. If a woman does not have breast cancer, the probability is 9.6% that she will also have a positive mammography. A woman in this age group had a positive mammography in a routine screening? What is the probability that she actually has breast cancer? A. greater than 90% B. between 70% and 90% C. between 50% and 70% D. between 30% and 50% E. between 10% and 30% F. less than 10% Is this typical or the exception? Perhaps high-level reasoning isn’t Bayesian but underlying mechanisms of learning, inference, memory, language, and perception are. 95 / 100 doctors correct answer WORK THROUGH BAYES RULE

Griffiths and Tenenbaum (2006) Optimal Predictions in Everyday Cognition If you were assessing an insurance case for an 18-year-old man, what would you predict for his lifespan? If you phoned a box office to book tickets and had been on hold for 3 minutes, what would you predict for the total time you would be on hold? If your friend read you her favorite line of poetry, and told you it was line 5 of a poem, what would you predict for the total length of the poem? If you opened a book about the history of ancient Egypt to a page listing the reigns of the pharaohs, and noticed that in 4000 BC a particular pharaoh had been ruling for 11 years, what would you predict for the total duration of his reign?

Griffiths and Tenenbaum Conclusion Average responses reveal a “close correspondence between peoples’ implicit probabilistic models and the statistics of the world.” People show a statistical sophistication and optimality of reasoning generally assumed to be absent in the domain of higher-order cognition.

Griffiths and Tenenbaum Bayesian Model If an individual has lived for tcur=50 years, how many years ttotal do you expect them to live?

What Does Optimality Entail? Individuals have complete, accurate knowledge about the domain priors. Fairly sophisticated computation involving Bayesian integral

From The Economist (1/5/2006) “[Griffiths and Tenenbuam]…put the idea of a Bayesian brain to a quotidian test. They found that it passed with flying colors.” “The key to successful Bayesian reasoning is … in having an appropriate prior… With the correct prior, even a single piece of data can be used to make meaningful Bayesian predictions.”

My Caution Bayesian formalism is sufficiently broad that nearly any theory can be cast in Bayesian terms E.g., adding two numbers as Bayesian inference Emphasis on how cognition conforms to Bayesian principles often directs attention away from important memory and processing limitations.

Latent Dirichlet Allocation (a.k.a. Topic Model) Problem Given a set of text documents, can we infer the topics that are covered by the set, and can we assign topics to individual documents Unsupervised learning problem Technique Exploit statistical regularities in data E.g., documents that are on the topic of education will likely contain a set of words such as ‘teacher’, ‘student’, ‘lesson’, etc.

Generative Model of Text Each document is a collection of topics (e.g., education, finance, the arts) Each topic is characterized by a set of words that are likely to appear The string of words in a document is generated by: Draw a topic from the probability distribution associated with a document Draw a word from the probability distribution associated with a topic Bag of words approach

Inferring (Learning) Topics Input: set of unlabeled documents Learning task Infer distribution over topics for each document Infer distribution over words for each topic Distribution over topics can be helpful for classifying or clustering documents

Topic Modeling Of Hotel Reviews Dan Knights, Rob Lindsey @ JD Powers

Phrase Discovery Rob Lindsey @ JD Powers

Value Of Probabilistic Models In AI And Cognitive Science AI and humans fundamentally have to deal with an uncertain (unknown) world Uncertainty is well described in the language of random events.

Value Of Probabilistic Models In AI And Cognitive Science Mathematically principled theories Currency of probability provides strong constraints on inference compare to neural net activations Probability is the optimal thing to compute, in the sense that any other strategy will lead to lower expected returns e.g., “I bet you $1 that roll of die will produce number < 3. How much are you willing to wager?”

Value Of Probabilistic Models In AI And Cognitive Science Elegant theories Key claims of theories are explicit Can minimize assumptions via Bayesian model averaging Often provides a direct means of incorporating prior knowledge

Value Of Probabilistic Models In AI And Cognitive Science Provides unified framework for describing pretty much any algorithm Allows you to see interrelationship among algorithms Allows you to develop new algorithms deterministic models are a subset of probabilistic models. neural nets are a subset of deterministic models

Important Technical Issues Representing structured data grammars relational schemas (e.g., paper authors, topics) Hierarchical models different levels of abstraction Nonparametric models flexible models that grow in complexity as the data justifies Inference: Exact vs. approximate Markov chain Monte Carlo, particle filters, variational approximations

Rationality in Cognitive Science Some theories in cognitive science are based on premise that human performance is optimal Rational theories, ideal observer theories Ignores biological constraints Probably true in some areas of cognition (e.g., vision) More interesting: bounded rationality Optimality is assumed to be subject to limitations on processing hardware and capacity, representation, experience with the world.