Building and evaluating models of human-level intelligence

Slides:



Advertisements
Similar presentations
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
Advertisements

The influence of domain priors on intervention strategy Neil Bramley.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Probabilistic Models in Human and Machine Intelligence.
Dynamic Bayesian Networks (DBNs)
Probabilistic Models of Cognition Conceptual Foundations Chater, Tenenbaum, & Yuille TICS, 10(7), (2006)
Induction. Inductive arguments increase the strength of your belief in some fact –Inductive arguments are not truth-preserving What sorts of factors should.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Bayesian models of inductive learning
Bayesian models of inductive learning
Part III Learning structured representations Hierarchical Bayesian models.
Bayesian models of human learning and reasoning Josh Tenenbaum MIT Department of Brain and Cognitive Sciences Computer Science and AI Lab (CSAIL)
Nonparametric Bayes and human cognition Tom Griffiths Department of Psychology Program in Cognitive Science University of California, Berkeley.
Part III Hierarchical Bayesian Models. Phrase structure Utterance Speech signal Grammar Universal Grammar Hierarchical phrase structure grammars (e.g.,
Exploring subjective probability distributions using Bayesian statistics Tom Griffiths Department of Psychology Cognitive Science Program University of.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Random Administrivia In CMC 306 on Monday for LISP lab.
Infinite block models for belief networks, social networks, and cultural knowledge Josh Tenenbaum, MIT 2007 MURI Review Meeting Work of Charles Kemp, Chris.
Modeling Vision as Bayesian Inference: Is it Worth the Effort? Alan L. Yuille. UCLA. Co-Director: Centre for Image and Vision Sciences. Dept. Statistics.
Normative models of human inductive inference Tom Griffiths Department of Psychology Cognitive Science Program University of California, Berkeley.
Statistical Natural Language Processing. What is NLP?  Natural Language Processing (NLP), or Computational Linguistics, is concerned with theoretical.
9.94 The cognitive science of intuitive theories J. Tenenbaum, T. Lombrozo, L. Schulz, R. Saxe.
Todo How to handle intro quote. Yet I’d like to think…. I’m not the only one. Make sure to practice talking through the shape bias part. –How to explain.
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
Bayesian models of human inductive learning Josh Tenenbaum MIT.
Framework for K-12 Science Education
Bayesian approaches to cognitive sciences. Word learning Bayesian property induction Theory-based causal inference.
Computer Science CPSC 322 Lecture 3 AI Applications 1.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Bayesian models of human inductive learning Josh Tenenbaum MIT Department of Brain and Cognitive Sciences Computer Science and AI Lab (CSAIL)
A Cognitive Substrate for Natural Language Understanding Nick Cassimatis Arthi Murugesan Magdalena Bugajska.
Introduction to probabilistic models of cognition Josh Tenenbaum MIT.
Bayesian models of inductive learning and reasoning Josh Tenenbaum MIT Department of Brain and Cognitive Sciences Computer Science and AI Lab (CSAIL)
Computational models of cognitive development: the grammar analogy Josh Tenenbaum MIT.
Bayesian models of inductive learning Tom Griffiths UC Berkeley Josh Tenenbaum MIT Charles Kemp CMU.
1 CS 2710, ISSP 2610 Foundations of Artificial Intelligence introduction.
Infinite block models for belief networks, social networks, and cultural knowledge Josh Tenenbaum, MIT 2007 MURI Review Meeting Work of Charles Kemp, Chris.
Definitions of AI There are as many definitions as there are practitioners. How would you define it? What is important for a system to be intelligent?
Probabilistic Models in Human and Machine Intelligence.
Motivation and Overview
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
1 Artificial Intelligence & Prolog Programming CSL 302.
Basic Bayes: model fitting, model selection, model averaging Josh Tenenbaum MIT.
Cognitive Modeling Cogs 4961, Cogs 6967 Psyc 4510 CSCI 4960 Mike Schoelles
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
Integrative Genomics I BME 230. Probabilistic Networks Incorporate uncertainty explicitly Capture sparseness of wiring Incorporate multiple kinds of data.
HTA as a framework for task analysis a review by Andrew Wright (2.8.16)
Sub-fields of computer science. Sub-fields of computer science.
Scientific Method.
Todo One slide on “beyond similarity-based induction”
Knowledge Representation Techniques
What is cognitive psychology?
Modeling human action understanding as inverse planning
A New Approach to Measure Preferences of Users in Built Environments: Integrating Cognitive Mapping and Utility Models Benedict Dellaert Erasmus University.
A I (Artificial Intelligence)
“Instead of trying to produce a programme to simulate the adult mind, why not rather try to produce one which simulates the child's?” Alan Turing, 1950.
Bayesian models of human learning and inference
Josh Tenenbaum Statistical learning of abstract knowledge:
Bayesian models of human learning and reasoning
Cognitive Language Comprehension in Rosie
Finding structure in data
Probabilistic Models in Human and Machine Intelligence
The Science of Biology! Chapter 1.
Statistical learning meets abstract knowledge: A challenge for the probabilistic-connectionist interface Josh Tenenbaum MIT.
The causal matrix: Learning the background knowledge that makes causal learning possible Josh Tenenbaum MIT Department of Brain and Cognitive Sciences.
AP Biology Class Bonneville High School
Introduction to Artificial Intelligence Instructor: Dr. Eduardo Urbina
Learning overhypotheses with hierarchical Bayesian models
Presentation transcript:

Building and evaluating models of human-level intelligence Josh Tenenbaum MIT Department of Brain and Cognitive Sciences Computer Science and AI Lab (CSAIL) Game plan: biology If 5 minutes left at the end, do words and theory acquisition. If no time left, just do theory acquisition.

The big problem of intelligence How do we go beyond the information given – draw inferences, make generalizations, and build models of the world from the sparse, noisy data of our experience? Ken: Human-scale learning includes: - Learning in small number of trials - Learning with rich, relational representations Nick: You can’t know whether you have a good computational model of how people solve some task unless you have a computational model that actually solves that task (i.e., AI and cog. sci. are not separable).

Word learning on planet Gazoob “tufa”

The big problem of intelligence The development of intuitive theories in childhood. Psychology: How do we learn to understand others’ actions in terms of beliefs, desires, plans, intentions, values, morals? Biology: How do we learn that people, dogs, bees, worms, trees, flowers, grass, coral, moss are alive, but chairs, cars, tricycles, computers, the sun, Roomba, robots, clocks, rocks are not?

The big problem of intelligence Common sense reasoning. Consider a man named Boris. Is the mother of Boris’s father his grandmother? Is the mother of Boris’s sister his mother? Is the son of Boris’s sister his son? (Note: Boris and his family were stranded on a desert island when he was a young boy.)

Approaches to cognitive modeling Micro approach: Build a processing model of a specific laboratory task. Pro: Can test precise behavioral predictions. Can establish clearly how the model works and why. Con: May not learn much of general value about the hardest, real-world problems of human cognition. model 1 model 2 … model n task 1 task 2 … task n

Approaches to cognitive modeling Macro approach: Build an architecture for modeling human cognition. Pro: Can build end-to-end human simulations. The same model applies to many tasks. Con: Hard to test. Many degrees of freedom, hard to know what is doing the work. Hard to export insights outside of the immediate modeling community. Cognitive architecture task 1 task 2 task n

Approaches to cognitive modeling Principle-based approach: Propose a small number of general-purpose principles for representation, learning, and inference; then use those principles to build models of many specific tasks. Examples: Connectionist, Bayesian, …. Pro: Unifying explanations of human cognition, supported by rigorously testable models. Principles supported by success of specific models they instantiate. Principles export well. Con: Principles not directly testable. No end-to-end human simulation. Principles model 1 model 2 … model n task 1 task 2 … task n

The “Bayesian” approach Probabilistic inference in generative models Hierarchical probabilistic models, with inference at all levels of abstraction Probabilities defined over structured representations: graphs, grammars, predicate logic, schemas Flexible representations, growing in complexity or changing form in response to the observed data. Approximate methods of learning and inference, such as Markov chain Monte Carlo (MCMC), to scale up to large problems.

Inductive reasoning Which argument is stronger? Gorillas have T9 hormones. Seals have T9 hormones. Anteaters have T9 hormones. “Similarity” “Typicality” “Diversity” Gorillas have T9 hormones. Seals have T9 hormones. Horses have T9 hormones. Gorillas have T9 hormones. Chimps have T9 hormones. Monkeys have T9 hormones. Baboons have T9 hormones. Horses have T9 hormones.

Beyond similarity-based induction Reasoning based on dimensional thresholds: (Smith et al., 1993) Reasoning based on causal relations: (Medin et al., 2004; Coley & Shafto, 2003) Poodles can bite through wire. German shepherds can bite through wire. Dobermans can bite through wire. German shepherds can bite through wire. Salmon carry E. Spirus bacteria. Grizzly bears carry E. Spirus bacteria. Grizzly bears carry E. Spirus bacteria. Salmon carry E. Spirus bacteria.

Theory-based Bayesian model (structural form + process) Property type “has T9 hormones” “can bite through wire” “carry E. Spirus bacteria” Theory-based Bayesian model (structural form + process) taxonomic tree directed chain directed network + diffusion process + drift process + noisy transmission Class D Class C Class G Class F Class E Class D Class B Class A Class D Class A Class A Class F Class E Class C Class C Class B Class G Class E Class B Class F Hypotheses Class G Class A Class B Class C Class D Class E Class F Class G . . . . . . . . .

“can bite through wire” “has T9 hormones” “can bite through wire” “carry E. Spirus bacteria” “is found near Minneapolis” (Kemp & Tenenbaum)

Integrating multiple forms of reasoning 2) Causal relations between features … Parameters of causal relations vary smoothly over the category hierarchy. 1) Taxonomic relations between categories T9 hormones cause elevated heart rates. Elevated heart rates cause faster metabolisms. Mice have T9 hormones. …? (Kemp, Shafto et al.)

Integrating multiple forms of reasoning

Learning domain structures P(structure | form) P(data | structure) P(form) F: form Tree with species at leaf nodes mouse squirrel chimp gorilla S: structure hormones Has T9 F1 F2 F3 F4 mouse squirrel chimp gorilla ? D: data … (Kemp & Tenenbaum)

Structural forms as graph grammars Process Form Process

Learning the abstract principles organizing a domain

Learning multiple structures within a domain (Shafto, Kemp, et al.)

Learning relational theories concept predicate concept Biomedical predicate data from UMLS (McCrae et al.): 134 concepts: enzyme, hormone, organ, disease, cell function ... 49 predicates: affects(hormone, organ), complicates(enzyme, cell function), treats(drug, disease), diagnoses(procedure, disease) … (Kemp, Tenenbaum, Griffiths et al.)

Learning relational theories e.g., Diseases affect Organisms Chemicals interact with Chemicals Chemicals cause Diseases

Learning abstract relational structures Dominance hierarchy Tree Cliques Ring Primate troop Bush administration Prison inmates Kula islands “x beats y” “x told y” “x likes y” “x trades with y”

Causal learning and reasoning Principles Structure Data (Griffiths, Tenenbaum, et al.)

The “blessing of abstraction” Task: learning a causal graphical model G 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 # of samples: 20 80 1000 Graph G edge (G) Data D edge (G) Abstract theory Z Graph G class (z) Data D

“Universal Grammar” Grammar Phrase structure Utterance Speech signal Hierarchical phrase structure grammars (e.g., CFG, HPSG, TAG) P(phrase structure | grammar) P(utterance | phrase structure) P(speech | utterance) (c.f. Chater and Manning, 2006) P(grammar | UG) Grammar Phrase structure Utterance Speech signal

Vision as probabilistic parsing (Han & Zhu, 2006; Yuille & Kersten, 2006 )

Goal-directed action (production and comprehension) (Wolpert et al., 2003)

Payoffs model 1 model 2 model n task 1 task 2 task n … Principles Ultimately, a framework for studying human knowledge: its nature, use, and acquisition, at multiple levels of abstraction, across domains and tasks. Export principles and insights to cognitive scientists outside the modeling community. Studying how principles interact takes us beyond traditional cog. sci. dichotomies (“Twenty questions”) A bridge to state-of-the-art AI. Future work: a bridge to neuroscience.