The Approach of Modern AI Stuart Russell Computer Science Division UC Berkeley.

Slides:



Advertisements
Similar presentations
Intelligent Agents Russell and Norvig: 2
Advertisements

Ai in game programming it university of copenhagen Reinforcement Learning [Outro] Marco Loog.
Introduction to Hierarchical Reinforcement Learning Jervis Pinto Slides adapted from Ron Parr (From ICML 2005 Rich Representations for Reinforcement Learning.
Class Project Due at end of finals week Essentially anything you want, so long as it’s AI related and I approve Any programming language you want In pairs.
ECE457 Applied Artificial Intelligence R. Khoury (2007)Page 1 Please pick up a copy of the course syllabus from the front desk.
AI 授課教師:顏士淨 2013/09/12 1. Part I & Part II 2  Part I Artificial Intelligence 1 Introduction 2 Intelligent Agents Part II Problem Solving 3 Solving Problems.
1 OSCAR: An Architecture for Generally Intelligent Agents John L. Pollock Philosophy and Cognitive Science University of Arizona
CPSC 322, Lecture 19Slide 1 Propositional Logic Intro, Syntax Computer Science cpsc322, Lecture 19 (Textbook Chpt ) February, 23, 2009.
LEARNING FROM OBSERVATIONS Yılmaz KILIÇASLAN. Definition Learning takes place as the agent observes its interactions with the world and its own decision-making.
Intelligence without Reason
Artificial Intelligence Overview John Paxton Montana State University August 14, 2003.
Introduction to Artificial Intelligence Prof. Kathleen McKeown 722 CEPSR, TAs: Kapil Thadani 724 CEPSR, Phong Pham TA Room.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Human-level AI Stuart Russell Computer Science Division UC Berkeley.
B. RAMAMURTHY EAP#2: Data Mining, Statistical Analysis and Predictive Analytics for Automotive Domain CSE651C, B. Ramamurthy 1 6/28/2014.
Reference: "Artificial Intelligence, a Modern Approach, 3rd ed."
Katanosh Morovat.   This concept is a formal approach for identifying the rules that encapsulate the structure, constraint, and control of the operation.
ARTIFICIAL INTELLIGENCE Introduction: Chapter 1. Outline Course overview What is AI? A brief history The state of the art.
1 AI and Agents CS 171/271 (Chapters 1 and 2) Some text and images in these slides were drawn from Russel & Norvig’s published material.
1 Lyle H. Ungar, University of Pennsylvania What is AI? “Artificial Intelligence is the study of how to make computers do things at which, at the moment,
Artificial Intelligence Dr. Paul Wagner Department of Computer Science University of Wisconsin – Eau Claire.
CISC4/681 Introduction to Artificial Intelligence1 Introduction – Artificial Intelligence a Modern Approach Russell and Norvig: 1.
Introduction: Chapter 1
AI Overview Reference: "Artificial Intelligence, a Modern Approach, 3 rd ed."
Ontologies Reasoning Components Agents Simulations Belief Update, Planning and the Fluent Calculus Jacques Robin.
Knowledge representation
IJCAI 2003 Workshop on Learning Statistical Models from Relational Data First-Order Probabilistic Models for Information Extraction Advisor: Hsin-His Chen.
Lection 3. Part 1 Chapter 2 of Russel S., Norvig P. Artificial Intelligence: Modern Approach.
Pattern-directed inference systems
Introduction to Artificial Intelligence and Soft Computing
Agents CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
1 CS 2710, ISSP 2610 Foundations of Artificial Intelligence introduction.
I Robot.
Artificial Intelligence Lecture 1. Objectives Definition Foundation of AI History of AI Agent Application of AI.
1 2010/2011 Semester 2 Introduction: Chapter 1 ARTIFICIAL INTELLIGENCE.
Simultaneously Learning and Filtering Juan F. Mancilla-Caceres CS498EA - Fall 2011 Some slides from Connecting Learning and Logic, Eyal Amir 2006.
BLOG: Probabilistic Models with Unknown Objects Brian Milch, Bhaskara Marthi, Stuart Russell, David Sontag, Daniel L. Ong, Andrey Kolobov University of.
Instructional Objective  Define an agent  Define an Intelligent agent  Define a Rational agent  Discuss different types of environment  Explain classes.
Rational Agency CSMC Introduction to Artificial Intelligence January 8, 2007.
What is Artificial Intelligence? What is artificial intelligence? It is the science and engineering of making intelligent machines, especially intelligent.
Definitions Think like humansThink rationally Act like humansAct rationally The science of making machines that: This slide deck courtesy of Dan Klein.
Rational Agency CSMC Introduction to Artificial Intelligence January 8, 2004.
What is Artificial Intelligence?
Feng Zhiyong Tianjin University Fall  An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that.
Intelligent Agents Introduction Rationality Nature of the Environment Structure of Agents Summary.
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
Introduction to Machine Learning © Roni Rosenfeld,
Generalized Point Based Value Iteration for Interactive POMDPs Prashant Doshi Dept. of Computer Science and AI Institute University of Georgia
CS382 Introduction to Artificial Intelligence Lecture 1: The Foundations of AI and Intelligent Agents 24 January 2012 Instructor: Kostas Bekris Computer.
Lecture 2: Intelligent Agents Heshaam Faili University of Tehran What is an intelligent agent? Structure of intelligent agents Environments.
From NARS to a Thinking Machine Pei Wang Temple University.
The Hebrew University of Jerusalem School of Engineering and Computer Science Academic Year: 2011/2012 Instructor: Jeff Rosenschein.
AI Overview Reference: "Artificial Intelligence, a Modern Approach, 3 rd ed."
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Artificial Intelligence
ECE457 Applied Artificial Intelligence R. Khoury (2007)Page 1 Please pick up a copy of the course syllabus from the front desk.
Introduction to Artificial Intelligence Prof. Kathleen McKeown 722 CEPSR Tas: Andrew Rosenberg Speech Lab, 7 th Floor CEPSR Sowmya Vishwanath TA Room.
Brief Intro to Machine Learning CS539
Introduction to Machine Learning
Artificial Intelligence
Basic Intro Tutorial on Machine Learning and Data Mining
Artificial Intelligence and Searching
Artificial Intelligence introduction(2)
Announcements Homework 3 due today (grace period through Friday)
AI and Agents CS 171/271 (Chapters 1 and 2)
Symbolic AI 2.0 Yi Zhou.
Expectation-Maximization & Belief Propagation
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Artificial Intelligence and Searching
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

The Approach of Modern AI Stuart Russell Computer Science Division UC Berkeley

Outline The History of AI: A Rational Reconstruction Unifying Formalisms Future Developments Structure in behaviour Cumulative learning Agent architectures Tasks and platforms Open problems

Early history

Early history contd. No integrated intelligent systems discipline emerged from the 1950s Different groups had different canonical problems and incompatible formalisms –Control theory: real vectors, linear/Gaussians –OR: atomic states, transition matrices –AI: first-order logic –Statistics: linear regression, Gaussian mixtures

Diaspora

Reunification(s)

Formalisms Communication among human mathematicians vs. abstract model of actual computational processes => focus on concrete syntax, semantics, completeness Multiple special-purpose models and algorithms vs. general representation and inference  can learn new things (e.g., multiplayer games) without qualitatively new representations and algorithms  need efficient special-case behavior to “fall out” from general methods applied to particular problems

[Milch, 2006]

Expressiveness Expressive language => concise models => fast learning, sometimes fast inference –E.g., rules of chess: 1 page in first-order logic, 100,000 pages in propositional logic, pages as atomic-state model Significant progress occurring, expanding contact layer between AI systems and real-world data

Expressiveness atomicpropositionalfirst-order/relational logic probability 5 th C B.C.19 th C 17 th C20 th C21 st C

Digression: GOFAI vs. Modern AI Technically, each logic is a special case of the corresponding probabilistic language Does probabilistic inference operate in the deterministic limit like logical inference? –One hopes so! –MCMC asymptotically identical to greedy SAT algorithms –Several algorithms exhibit “default logic” phenomenology Modern AI should build on, not discard, GOFAI

First-order probabilistic languages Gaifman [1964]: –distributions over first-order possible worlds Halpern [1990]: –syntax for constraints on such distributions Poole [1993] and several others: –KB defines distribution exactly –assumes unique names and domain closure Milch et al [2005] and others: –distributions over full first-order possible worlds generative models for events and existence –complete inference for all well-defined models

Possible worlds Propositional First-order + unique names, domain closure First-order open-world A B C D

Why does this matter? What objects are referred to in the following sentence?

Why does this matter? What objects appear in this image?

Unifications Bayes nets [Pearl, 1988] led to the integration of (parts of) machine learning, statistics, KR, expert systems, speech recognition First-order probabilistic languages may help to unify the rest of ML and KR, information extraction, NLP, vision, tracking and data association

Structure in behaviour One billion seconds, one trillion (parallel) actions Unlikely to be generated from a flat solution to the unknown POMDP of life Hierarchical structuring of behaviour: enunciating this syllable saying this word saying this sentence explaining structure in behaviour giving a talk about AI …….

Hierarchical reinforcement learning Hierarchical structure expressed as partial program – declarative procedural knowledge: what (not) to do Key point: decisions within “subroutines” are independent of almost all state variables –E.g., say(word,prosody) not say(word,prosody,NYSEprices,NASDAQ…) Value functions decompose into additive factors, both temporally and functionally => fast learning Decisions depend on internal + external state! –E.g., I don’t even consider selling IBM stock during talk

HRL contd. Complete learning algorithms exist for general-purpose, concurrent partial programming languages, but... We need a KR subfield for “know-how” Where do subroutines come from? More generally, how are useful composites (objects, actions) built up?

Learning data Learning knowledge

Learning data Learning knowledge prior knowledge

Learning data Learning knowledge prior knowledge

Learning data Learning knowledge prior knowledge Learned knowledge in a form usable as prior knowledge, not always “one step down” in abstraction Self-reinforcing accumulation of knowledge and representational abstractions

Agent architectures Core: state estimation maintains belief state –Lots of progress; extend to open worlds Reflex: π(s) Action-value: argmax a Q(s,a) Goal-based: a such that G(result(a,s)) Utility-based: argmax a E(U(result(a,s))) Would be nice to understand when one is better (e.g., more learnable) than another Need an agent-building language providing composition and decomposition of architectural components

Metareasoning Computations are actions too Controlling them is essential, esp. for model-based architectures that can do lookahead and for approximate inference on intractable models Effective control based on expected value of computation Methods for learning this must be built-in – brains unlikely to have fixed, highly engineered algorithms that will correctly dictate trillions of computational actions over agent’s lifetime

A science of architecture My boxes and arrows vs your boxes and arrows? Well-designed/evolved architecture solves what optimization problem? What forces drive design choices? –Generate optimal actions –Generate them quickly –Learn to do this from few experiences Each by itself leads to less interesting solutions –omitting learnability favours the all-compiled solution –omitting resource bounds favours Bayes-on-Turing-machines Bounded-optimal solutions have interesting architectures and can be found for some nontrivial cases

Challenge problems should involve… Continued existence Behavioral structure at several time scales (not just repetition of small task) Finding good decisions should sometimes require extended deliberation Environment with many, varied objects, nontrivial perception (other agents, language optional) Examples: cook, house cleaner, secretary, courier Wanted: a human-scale dextrous 4-legged robot + simulation-based CAD for vision-based robotics

Open problems Learning better representations – need a new understanding of reification/generalization –Why did Eurisko’s nested loops run out of gas? Learning new behavioural structures Generating new goals from utility soup Do neuroscience and cognitive science have anything to tell us? What if we succeed? Can we design probably approximately “safe” agents?