Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Cognitive Architectures and Virtual Intelligent Agents Thanks.

Slides:



Advertisements
Similar presentations
Problem- Based Learning in STEM Disciplines Saturday, November 10, 2007 JHU/MSU STEM Initiative.
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Arizona State University and Institute for the Study of Learning and Expertise Expertise, Transfer, and Innovation in.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California Elena Messina.
Center for the Study of Language and Information Stanford University, Stanford, California March 20-21, 2004 Symposium on Reasoning and Learning in Cognitive.
Pat Langley Seth Rogers Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA.
Pat Langley Institute for the Study of Learning and Expertise Palo Alto, CA Cumulative Learning of Relational and Hierarchical Skills.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Varieties of Problem Solving in a Unified Cognitive Architecture.
Pat Langley Dan Shapiro Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Extending the I CARUS Cognitive Architecture Thanks to D. Choi,
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Pat Langley Dongkyu Choi Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Mental Simulation and Learning in the I CARUS Architecture.
General learning in multiple domains transfer of learning across domains Generality and Transfer in Learning training items test items training items test.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Modeling Social Cognition in a Unified Cognitive Architecture.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Information Processing Technology Office Learning Workshop April 12, 2004 Seedling Overview Learning Hierarchical Reactive Skills from Reasoning and Experience.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
IL Kickoff Meeting June 20-21, 2006 DARPA Integrated Learning POIROT Project 1 Learning Hierarchical Task Networks by Analyzing Expert Traces Pat Langley.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona Institute for the Study of Learning and Expertise Palo Alto, California.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Institute for the Study of Learning and Expertise Palo Alto,
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona A Cognitive Architecture for Integrated.
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA A Unified Cognitive Architecture for Embodied Agents Thanks.
Pat Langley Computer Science and Engineering Arizona State University Tempe, Arizona USA Institute for the Study of Learning and Expertise Palo Alto, California.
Pat Langley Institute for the Study of Learning and Expertise Palo Alto, California A Cognitive Architecture for Complex Learning.
1 ISLE Transfer Learning Team Main Technology Components The I CARUS Architecture Markov Logic Networks Executes skills in the environment Long-TermConceptualMemoryShort-TermConceptualMemory.
Cognitive Systems, ICANN panel, Q1 What is machine intelligence, as beyond pattern matching, classification and prediction. What is machine intelligence,
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Threaded Cognition: An Integrated Theory of Concurrent Multitasking
Introduction to SOAR Based on “a gentle introduction to soar: an Architecture for Human Cognition” by Jill Fain Lehman, John Laird, Paul Rosenbloom. Presented.
TRANSIMS Research and Deployment Project TRACC TSM Staff Dr. Vadim Sokolov Dr. Joshua Auld Dr. Kuilin Zhang Mr. Michael Hope.
SPECIFYING COGNITIVE MODELS Using Patterns and Conflicts A. Macklem, F. Mili Oakland University S. Dungrani TARDEC June, 2004.
The Importance of Architecture for Achieving Human-level AI John Laird University of Michigan June 17, th Soar Workshop
Polyscheme John Laird February 21, Major Observations Polyscheme is a FRAMEWORK not an architecture – Explicitly does not commit to specific primitives.
Models of Human Performance Dr. Chris Baber. 2 Objectives Introduce theory-based models for predicting human performance Introduce competence-based models.
Abstraction and ACT-R. Outline Motivations for the Theory –Architecture –Abstraction Production Rules ACT-R Architecture of Cognition.
Meaningful Learning in an Information Age
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
Knowledge representation
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
SOFTWARE DESIGN.
Synthetic Cognitive Agent Situational Awareness Components Sanford T. Freedman and Julie A. Adams Department of Electrical Engineering and Computer Science.
By: Chantae Garrett, Michelle Cairy, Kenneth Guidry, and Raul Garcia.
Artificial Intelligence By Michelle Witcofsky And Evan Flanagan.
1 CS 2710, ISSP 2610 Foundations of Artificial Intelligence introduction.
1 Module F1 Modular Training Cycle and Integrated Curriculum.
Korea Univ. Division Information Management Engineering UI Lab. Korea Univ. Division Information Management Engineering UI Lab. IMS 802 Cognitive Modeling.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Cognitive Science and Biomedical Informatics Department of Computer Sciences ALMAAREFA COLLEGES.
Cognitive Architectures For Physical Agents Sara Bolduc Smith College CSC 290.
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Vedrana Vidulin Jožef Stefan Institute, Ljubljana, Slovenia
Cognitive Architectures and General Intelligent Systems Pay Langley 2006 Presentation : Suwang Jang.
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, CA
Artificial Intelligence
Air Force Institute of Technology
Learning Fast and Slow John E. Laird
Artificial intelligence (AI)
CHAPTER 1 Introduction BIC 3337 EXPERT SYSTEM.
Learning Teleoreactive Logic Programs by Observation
Model-Driven Analysis Frameworks for Embedded Systems
MANAGING KNOWLEDGE FOR THE DIGITAL FIRM
Introduction Artificial Intelligent.
Symbolic cognitive architectures
Artificial Intelligence Lecture 2: Foundation of Artificial Intelligence By: Nur Uddin, Ph.D.
EA C461 – Artificial Intelligence Introduction
A Value-Driven Architecture for Intelligent Behavior
Presentation transcript:

Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Cognitive Architectures and Virtual Intelligent Agents Thanks to D. Choi, G. Cleveland, T. Konik, N. Li, N. Nejati, C. Park, and D. Stracuzzi for their contributions. This talk reports research partly funded by grants from ONR and DARPA, which are not responsible for its contents.

The Nature of Intelligence Cognitive science and AI, although distinct fields, agree on some core assumptions: Intelligence arises from computational processes A variety of functional abilities underlie intelligence These abilities are general yet benefit from knowledge Humans remain our best examples of intelligent systems Progress toward understanding intelligence requires taking a position on two key methodological questions: What theoretical framework should one assume? What problems or testbeds should drive research?

Frameworks for Intelligent Systems Because intelligence involves distinct abilities, it is tempting to develop them separately and then combine them. This position is well represented in the AI community by: Software engineering (e.g., Brugali, 2009) Blackboard systems (Engelmore & Morgan, 1988) Multi-agent systems (Weiss, 2000) Such integrated approaches offer clear advantages in terms of modular development and division of labor. But they are not the only way to create intelligent artifacts.

Cognitive Architectures A cognitive architecture (Newell, 1990) is an infrastructure for intelligent systems that: makes strong theoretical assumptions about the representations and mechanisms underlying cognition incorporates many ideas from psychology about the nature of the human mind contains distinct modules, but these access and alter the same memories and representations comes with a programming language that eases construction of knowledge-based systems A cognitive architecture is all about mutual constraints, and it should provide a unified theory of intelligent behavior.

Testbeds for Intelligent Systems We also need problems and testbeds that pose challenges to theories and evaluate their responses. Examples include: Educational tasks (e.g., reasoning in math and physics) Standardized tests (e.g., IQ, SAT, GRE exams) Integrated robots (e.g., for search and rescue) Synthetic characters in virtual worlds (e.g., for games) Each problem class has advantages and disadvantages, but all have important roles to play. In this talk, I will report work on synthetic characters that has driven our progress on cognitive architectures.

Synthetic Agents for Urban Driving Insert demo here

Outline for the Talk Frameworks and testbeds for intelligent systems Review of the I CARUS cognitive architecture Common and distinctive features Representation and organization of memory Performance and learning mechanisms Synthetic agents developed with I CARUS Complex synthetic environments Lessons learned from these efforts Plans for future architectural / testbed research

The I CARUS Architecture I CARUS (Langley, 2006) is a computational theory of the human cognitive architecture that posits: These assumptions are not novel; it shares them with architectures like Soar (Laird et al., 1987) and ACT-R (Anderson, 1993) Short-term memories are distinct from long-term stores 2. 2.Memories contain modular elements cast as symbolic structures 3. 3.Long-term structures are accessed through pattern matching 4. 4.Cognition occurs in retrieval/selection/action cycles 5. 5.Learning involves monotonic addition of elements to memory 6. 6.Learning is incremental and interleaved with performance

Distinctive Features of I CARUS However, I CARUS also makes assumptions that distinguish it from these architectures: Some of these tenets also appear in Bonasso et al.s (2003) 3T, Freeds (1998) APEX, and Sun et al.s (2001) Clarion. But only I CARUS combines them into a unified cognitive theory Cognition is grounded in perception and action 2. 2.Categories and skills are separate cognitive entities 3. 3.Short-term elements are instances of long-term structures 4. 4.Skills and concepts are organized in a hierarchical manner 5. 5.Inference and execution are more basic than problem solving

Research Goals for I CARUS Our current objectives in developing I CARUS are to produce: a computational theory of high-level cognition in humans that is qualitatively consistent with results from psychology that exhibits as many distinct cognitive functions as possible that supports creation of intelligent agents in virtual worlds We have not yet modeled quantitative results from experiments, but we hope to obtain such fits when I CARUS is more mature. This would enable more direct comparison to architectures like ACT-R, Clarion, and EPIC (Meyer & Kieras, 1997).

Cascaded Integration in I CARUS I CARUS adopts a cascaded approach to integration in which lower-level modules produce results for higher-level ones. conceptual inference skill execution problem solving learning Like other unified cognitive architectures, I CARUS incorporates a number of distinct modules.

I CARUS Beliefs and Goals for Urban Driving Inferred beliefs: (current-street me A)(current-segment me g550) (lane-to-right g599 g601)(first-lane g599) (last-lane g599)(last-lane g601) (under-speed-limit me)(slow-for-right-turn me) (steering-wheel-not-straight me)(centered-in-lane me g550 g599) (in-lane me g599)(in-segment me g550) (on-right-side-in-segment me)(intersection-behind g550 g522) (building-on-left g288)(building-on-left g425) (building-on-left g427)(building-on-left g429) (building-on-left g431)(building-on-left g433) (building-on-right g287)(building-on-right g279) (increasing-direction me)(near-pedestrian me g567) Top-level goals: (not (near-pedestrian me ?any))(not (near-vehicle me ?other)) (on-right-side-in-segment me)(in-lane me ?segment) (not (over-speed-limit me))(not (running-red-light me))

I CARUS Concepts for Urban Driving ((in-rightmost-lane ?self ?clane) :percepts ((self ?self) (segment ?seg) (line ?clane segment ?seg)) :relations ((driving-well-in-segment ?self ?seg ?clane) (last-lane ?clane) (not (lane-to-right ?clane ?anylane)))) ((driving-well-in-segment ?self ?seg ?lane) :percepts ((self ?self) (segment ?seg) (line ?lane segment ?seg)) :relations ((in-segment ?self ?seg) (in-lane ?self ?lane) (aligned-with-lane-in-segment ?self ?seg ?lane) (centered-in-lane ?self ?seg ?lane) (steering-wheel-straight ?self))) ((in-lane ?self ?lane) :percepts ((self ?self segment ?seg) (line ?lane segment ?seg dist ?dist)) :tests ((> ?dist -10) (<= ?dist 0)))

Hierarchical Organization of Concepts I CARUS organizes conceptual memory in a hierarchical manner. The same conceptual predicate can appear in multiple clauses to specify disjunctive and recursive concepts. concept concept clause percept

Conceptual Inference in I CARUS Conceptual inference in I CARUS occurs from the bottom up. Starting with observed percepts, this process produces high-level beliefs about the current state. concept concept clause percept

Conceptual Inference in I CARUS Conceptual inference in I CARUS occurs from the bottom up. Starting with observed percepts, this process produces high-level beliefs about the current state. concept concept clause percept

Conceptual Inference in I CARUS Conceptual inference in I CARUS occurs from the bottom up. Starting with observed percepts, this process produces high-level beliefs about the current state. concept concept clause percept

Conceptual Inference in I CARUS Conceptual inference in I CARUS occurs from the bottom up. Starting with observed percepts, this process produces high-level beliefs about the current state. concept concept clause percept

((in-rightmost-lane ?self ?line) :percepts ((self ?self) (line ?line)) :start ((last-lane ?line)) :subgoals ((driving-well-in-segment ?self ?seg ?line))) ((driving-well-in-segment ?self ?seg ?line) :percepts ((segment ?seg) (line ?line) (self ?self)) :start ((steering-wheel-straight ?self)) :subgoals ((in-segment ?self ?seg) (centered-in-lane ?self ?seg ?line) (aligned-with-lane-in-segment ?self ?seg ?line) (steering-wheel-straight ?self))) ((in-segment ?self ?endsg) :percepts ((self ?self speed ?speed) (intersection ?int cross ?cross) (segment ?endsg street ?cross angle ?angle)) :start ((in-intersection-for-right-turn ?self ?int)) :actions ((steer 1))) I CARUS Skills for Urban Driving

Hierarchical Organization of Skills I CARUS organizes skills in a hierarchical manner, which each skill clause indexed by the goal it aims to achieve. goal skill clause operator The same goal can index multiple clauses to allow disjunctive, conditional, and recursive procedures.

Skill Execution in I CARUS A skill clause is applicable if its goal is unsatisfied and if its conditions hold, given bindings from above. Skill execution occurs from the top down, starting from goals, to find applicable paths through the skill hierarchy. goal skill clause operator

Skill Execution in I CARUS However, I CARUS prefers to continue ongoing skills when they match, giving it a bias toward persistence over reactivity. This process repeats on each cycle to produce goal-directed but reactive behavior (Nilsson, 1994). goal skill clause operator

Skill Execution in I CARUS At this point, another unsatisfied goal begins to drive behavior, invoking different skills to pursue it. If events proceed as expected, this iterative process eventually achieves the agents top-level goal. goal skill clause operator

Execution and Problem Solving in I CARUS Skill Hierarchy Reactive Execution impasse? Problem Solving yes no Problem solving involves means-ends analysis that chains backward over skills and concept definitions, executing skills whenever they become applicable. Primitive Skills Problem goal beliefs Generated Plan

I CARUS Learns Skills from Problem Solving Skill Hierarchy Reactive Execution impasse? Problem Solving no Primitive Skills Problem goal beliefs Generated Plan Skill Learning yes

Learning from Problem Solutions operates whenever problem solving overcomes an impasse incorporates only information stored locally with goals generalizes beyond the specific objects concerned depends on whether chaining involved skills or concepts supports cumulative learning and within-problem transfer I CARUS incorporates a mechanism for learning new skills that: This skill creation process is fully interleaved with means-ends analysis and execution. Learned skills carry out forward execution in the environment rather than backward chaining in the mind.

I CARUS Summary includes hierarchical memories for concepts and skills; interleaves conceptual inference with reactive execution; resorts to problem solving when it lacks routine skills; learns such skills from successful resolution of impasses. I CARUS is a unified theory of the cognitive architecture that: We have developed I CARUS agents for a variety of simulated physical environments. However, each effort has revealed limitations that have led to important architectural extensions.

Creating Synthetic Agents in I CARUS Urban Combat Urban Driving Rush 2008 MadRTS

Urban Combat is a synthetic environment, built on the Quake engine, used in the DARPA Transfer Learning program. Tasks for agents involved traversing an urban landscape with a variety of obstacles to capture a flag. Each of these insights has influenced our more recent work on I CARUS. Synthetic Agents for Urban Combat Storing, using, and learning route knowledge Learning to handle obstacles through trial and error Supporting different varieties of structural transfer across problems Our experiences in Urban Combat provided some clear lessons:

Source IED Start IED Start IED Target Transfer Studies with Urban Combat Transfer vs. Control Conditions Experiments with I CARUS agents on Urban Combat demonstrated varieties of structural transfer across multiple goal-directed tasks (Choi et al., ICCM-2007).

Synthetic Agents for Urban Driving We have developed an urban driving environment using Garage Games Torque game engine. This domain involves a mixture of cognitive and sensori-motor behavior in a constrained but complex and dynamic setting.

The Torque urban driving environment supports tasks of graded complexity, including: Keeping the car aligned at a constant speed Changing speed to obey traffic signs and lights Changing lanes/speed to avoid vehicles/pedestrians Approaching and turning a corner Driving repeatedly around a block Delivering packages to specified addresses We have developed I CARUS agents that address each of these tasks in a reasonably robust manner. We have also demonstrated its ability to learn driving skills from problem solving (Langley & Choi, JMLR, 2006). Synthetic Agents for Urban Driving

Our early efforts on urban driving motivated two key features of the I CARUS architecture: Indexing skills by the goals they achieve Multiple top-level goals with priorities Our use of this complex environment has driven our cognitive architecture research toward important new functionalities. Synthetic Agents for Urban Driving Long-term memory for generic goals Reactive generation of specific short-term goals Resource-based execution of multiple skills per cycle Recent work on Torque driving agents has led to additional changes to the architecture (Choi, 2010):

We have also developed I CARUS agents to execute football plays in Rush 2008 from Knexus Research. This domain is less complex in some ways but it still remains highly reactive and depends on spatio-temporal coordination among players on a team. Synthetic Agents for American Football

Our results in Rush 2008 are interesting because I CARUS learned its football skills by observing other agents. This and similar videos of the Oregon State football team provided training examples to drive this process. Synthetic Agents for American Football

Time stamps on beliefs to support episodic traces Ability to represent and recognize temporal concepts Adapting means-ends analysis to explain observed behavior Ability to control multiple agents in an environment Synthetic Agents for American Football We have used the extended I CARUS to learn hierarchical skills for 20 distinct football plays (Li et al., AAIDE-2009). These results utilized Oregon States image-processing system, which extracted objects and simple events from the videos. Our efforts on Rush 2008 have motivated additional extensions to the I CARUS architecture:

Synthetic Agents for Twig Scenarios Most recently, we have used Horswills (2008) Twig simulator to develop humanoid I CARUS agents. This low-fidelity environment supports a few object types, along with simple reactive behaviors for virtual characters.

We have developed a specific Twig scenario that involves four types of synthetic characters: Innocents go from ball to ball, pausing to pick up any dolls they encounter nearby. Collectors sit in chairs, buy dolls from anyone who offers at the right price, and secure these purchases in safes. Producers stand by trees, produce new dolls, and guard them by blocking any other agents who approach. Capitalists work in pairs, buying or stealing dolls from Innocents, stealing them from Producers, and selling them to Collectors. We have developed I CARUS agents for these character types to demonstrate their interactions with each other. This served as the final project for an ASU course on cognitive systems and intelligent agents ( Synthetic Agents for Twig Scenarios

Yet humans can reason about the beliefs, goals, and intentions of others, then use their inferences to make decisions. Reasoning about Others The framework can deal with other agents, but only by viewing them as objects in the environment. We designed I CARUS to model intelligent embodied agents, but our work has emphasized independent action. One might even argue that such deep social cognition is a key feature of human intelligence. Adding this capability to I CARUS will require extensions to its representation, performance processes, and learning methods.

Ongoing I CARUS Extensions To provide I CARUS with the capability to reason about others mental states, we are: Extending its representation to allow embedded modal literals [e.g., (belief me (goal driver2 (in-right-lane driver2 seg6)))] ; Revising inference to support incremental abductive reasoning that explains events in terms of mental states; Extending skill execution and problem solving to pursue goals that change mental states (e.g., by communication); and Introducing learning mechanisms that make inference about others more efficient with experience. We are developing these abilities to construct conversational agents for emergency medical assistance in the field.

Synthetic Environments / Architectural Testbeds In an STTR collaboration, we are using SET Corp.s CASTLE, a flexible physics/simulation engine, to: Create an improved urban driving environment that supports a richer set of objects, goals, and activities; Develop a treasure hunt testbed in which players must follow complex instructions and coordinate with other agents; Create interfaces to I CARUS, ACT-R, and other architectures; Design tasks of graded complexity that can aid in evaluating architecture-based cognitive models. Together, these will let the community evaluate its architectures on common problems in shared testbeds. These will support both extended, open-ended operations and brief, hand-crafted scenarios.

Rigid body dynamics solid objects, Newtonian physics, friction, joints, springs; no flexible objects or fluids Z-buffer graphics solid objects, textures, lights; no shadows, motion blur, or irradiance CASTLE Screen Shots / Assumptions

Concluding Remarks Grounds high-level cognition in perception and action Treats categories and skills as separate cognitive entities Views short-term structures as instances of long-term ones Organizes skills and concepts in a hierarchical manner Combines reactive control with deliberative problem solving In this talk, I presented I CARUS, a unified theory of the human cognitive architecture that: In addition, I reported a number of I CARUS agents that control synthetic agents in simulated environments. These efforts raised theoretical challenges, led to extensions, and revealed insights about the nature of intelligence.

Other Research on Modeling / Simulation – + + – Time – LysosomeCytoplasm Fe ROS lytic-enzyme membrane-damage lipofuscin junk-proteinjunk-protein H202 lipofuscin H202 oxidized-protein Quantitative process models of ecological systems (Bridewell et al., MLj, 2008) Qualitative system-level models of human aging (Langley, NIA-2009)

End of Presentation