Summarized by Eun Seok Lee BioIntelligence Lab 20 Sep, 2012

Slides:



Advertisements
Similar presentations
Pat Langley School of Computing and Informatics Arizona State University Tempe, Arizona USA Modeling Social Cognition in a Unified Cognitive Architecture.
Advertisements

The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel
The influence of domain priors on intervention strategy Neil Bramley.
Chapter 2: Marr’s theory of vision. Cognitive Science  José Luis Bermúdez / Cambridge University Press 2010 Overview Introduce Marr’s distinction between.
Bayesian models for fMRI data
The free-energy principle: a rough guide to the brain? Karl Friston
Yiannis Demiris and Anthony Dearden By James Gilbert.
Introduction to Neural Network Justin Jansen December 9 th 2002.
December 1, 2009Introduction to Cognitive Science Lecture 22: Neural Models of Mental Processes 1 Some YouTube movies: The Neocognitron Part I:
Baysian Approaches Kun Guo, PhD Reader in Cognitive Neuroscience School of Psychology University of Lincoln Quantitative Methods 2011.
Vision in Man and Machine. STATS 19 SEM Talk 2. Alan L. Yuille. UCLA. Dept. Statistics and Psychology.
ART AND PERCEPTUAL EXPERTISE. Structure Applying principles of perception to art-making Why study representational drawing? How to quantify drawing ability.
Science and Engineering Practices
How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to understand.
Biointelligence Laboratory School of Computer Science and Engineering Seoul National University Cognitive Robots © 2014, SNU CSE Biointelligence Lab.,
Free energy and active inference Karl Friston Abstract How much about our interaction with – and experience of – our world can be deduced from basic principles?
Optimality in Motor Control By : Shahab Vahdat Seminar of Human Motor Control Spring 2007.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Machine Learning. Learning agent Any other agent.
Abstract We start with a statistical formulation of Helmholtz’s ideas about neural energy to furnish a model of perceptual inference and learning that.
The free-energy principle: a rough guide to the brain? K Friston Summarized by Joon Shik Kim (Thu) Computational Models of Intelligence.
An Architecture for Empathic Agents. Abstract Architecture Planning + Coping Deliberated Actions Agent in the World Body Speech Facial expressions Effectors.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Learning Science and Mathematics Concepts, Models, Representations and Talk Colleen Megowan.
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
Varieties of Helmholtz Machine Peter Dayan and Geoffrey E. Hinton, Neural Networks, Vol. 9, No. 8, pp , 1996.
A New Theory of Neocortex and Its Implications for Machine Intelligence TTI/Vanguard, All that Data February 9, 2005 Jeff Hawkins Director The Redwood.
DCM for ERPs/EFPs Clare Palmer & Elina Jacobs Expert: Dimitris Pinotsis.
Graphical Models in Vision. Alan L. Yuille. UCLA. Dept. Statistics.
Abstract This talk summarizes recent attempts to integrate action and perception within a single optimization framework. We start with a statistical formulation.
Methodology Matters: Doing Research in the Behavioral and Social Sciences ICS 205 Ha Nguyen Chad Ata.
MURI: Integrated Fusion, Performance Prediction, and Sensor Management for Automatic Target Exploitation 1 Dynamic Sensor Resource Management for ATE MURI.
Autonomy and Artificiality Margaret A. Boden Hojin Youn.
Information: Perception and Representation Lecture #7 Part A.
Abstract This talk summarizes our recent attempts to integrate action and perception within a single optimization framework. We start with a statistical.
Abstract We offer a formal treatment of choice behaviour based on the premise that agents minimise the expected free energy of future outcomes. Crucially,
Abstract Predictive coding models and the free-energy principle, suggests that cortical activity in sensory brain areas reflects the precision of prediction.
Abstract This presentation questions the need for reinforcement learning and related paradigms from machine-learning, when trying to optimise the behavior.
Abstract How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts.
How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to understand.
Free energy and active inference
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
Chapter 2. From Complex Networks to Intelligent Systems in Creating Brain-like Systems, Sendhoff et al. Course: Robots Learning from Humans Baek, Da Som.
Cognitive Modular Neural Architecture
Abstract We suggested recently that attention can be understood as inferring the level of uncertainty or precision during hierarchical perception. In.
Abstract This presentation will look at action, perception and cognition as emergent phenomena under a unifying perspective: This Helmholtzian perspective.
Workshop on Mathematical Models of Cognitive Architectures December 5-9, 2011 CIRM, Marseille Workshop on Mathematical Models of Cognitive Architectures.
How much about our interaction with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to understand.
Grounded cognition. Barsalou, L. W. (2008). Annual Review of Psychology, 59, Grounded theories versus amodal representations. – Recapitulation.
Chapter 2. From Complex Networks to Intelligent Systems in Creating Brain-Like Intelligence, Olaf Sporns Course: Robots Learning from Humans Park, John.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Dynamic Causal Model for evoked responses in MEG/EEG Rosalyn Moran.
Chapter 15. Cognitive Adequacy in Brain- Like Intelligence in Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans Cinarel, Ceyda.
Chapter 4. Analysis of Brain-Like Structures and Dynamics (2/2) Creating Brain-Like Intelligence, Sendhoff et al. Course: Robots Learning from Humans 09/25.
Abstract How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Cognitive Modeling Cogs 4961, Cogs 6967 Psyc 4510 CSCI 4960 Mike Schoelles
Bayesian Brain - Chapter 11 Neural Models of Bayesian Belief Propagation Rajesh P.N. Rao Summary by B.-H. Kim Biointelligence Lab School of.
Bayesian Model Selection and Averaging SPM for MEG/EEG course Peter Zeidman 17 th May 2016, 16:15-17:00.
A probabilistic approach to cognition
Mihály Bányai, Vaibhav Diwadkar and Péter Érdi
Nicolas Alzetta CoNGA: Cognition and Neuroscience Group of Antwerp
Neural mechanisms underlying repetition suppression in occipitotemporal cortex Michael Ewbank MRC Cognition and Brain Sciences Unit, Cambridge, UK.
The User Lecture 2 DeSiaMore
Computational models for imaging analyses
Wellcome Trust Centre for Neuroimaging University College London
The free-energy principle: a rough guide to the brain? K Friston
CHAPTER I. of EVOLUTIONARY ROBOTICS Stefano Nolfi and Dario Floreano
Presentation transcript:

Summarized by Eun Seok Lee BioIntelligence Lab 20 Sep, 2012 Whatever Next? Predictive Brains, Situated Agents, and the Future of Cognitive Science Andy Clark Summarized by Eun Seok Lee BioIntelligence Lab 20 Sep, 2012

Abstract “Brains are essentially prediction machines.” Brains support perception and action by constantly attempting to match incoming sensory inputs with top-down expectations or predictions. A hierarchical generative model – Minimize prediction error A unifying model of perception and action A ‘hierarchical prediction machine’ approach – It offers the best clue yet to the shape of a unified science of mind and action Sections 1 and 2 – Key elements and implications of the approach Section 3 – Evidential, methodological, and the more properly conceptual pitfalls and challenges Sections 4 and 5 – Impact of such approaches on our more general vision of mind, experience, and agency

Contents 1. Prediction Machines 2. Representation, Inference, and the Continuity of Perception, Cognition, and Action 3. From Action-Oriented Predictive Processing to an Architecture of Mind. 4. Content and Consciousness 5. Taking Stock (Discussion & Conclusion)

1. Prediction Machines 1.1 From Helmholtz to Action-Oriented Predictive Processing 1.2 Escaping the Black Box 1.3 Dynamic Predictive Coding by the Retina 1.4 Another Illustration: Binocular Rivalry 1.5 Action-Oriented Predictive Processing 1.6 The Free Energy Formulation

The (Traditional) Architecture of Mind and Action Top-down Bottom-up Self-Consciousness Explicit Learning Inference Concept Formaton Action Recognition Memoization Action Selection Attention Implicit Learning Perception Mind Architecture (traditional)

Analogies of Mind and Next One (1/4) 16-17 C.

Analogies of Mind and Next One (2/4) 18-19 C.

Analogies of Mind and Next One (3/4) 20 C.

Analogies of Mind and Next One (4/4) 20 C. ~

From Helmholtz to Action-Oriented Predictive Processing The whole function of the brain is summed up in: error correction – Ross Ashby – “How to minimize errors?” Helmholtz (1860) in depicting perception as a process of probabilistic, knowledge-driven inference Analysis by Synthesis – brain tries to predict the current suite of cues from its best models of the possible causes Helmholtz Machine and its tradition – ‘back propagation’ and ‘Helmholtz Machine’ – learn new representations in a multi-level system (thus capturing increasingly deep regularities within a domain) without requiring the provision of copious preclassified samples of the desired input-output mapping (see Hinton (2007a)) Predictive coding – depicts the top-down flow as attempting to predict and fully ‘explain away’ the driving sensory signal, leaving only any residual ‘prediction errors’ to propagate information forward within the system

Evidence: Binocular Rivalry (1/2)

Evidence: Binocular Rivalry (2/2) Incoming signals remain constant while the percept switches to and fro (Frith, Perry, and Lumer (1999)) Hierarchical generative model -- explain away the incoming sensory signal by means of a matching top-down prediction ‘Empirical Bayes’ – the higher level guesses are thus acting as priors for the lower level processing “Makes the best predictions and that, taking priors into consideration, is consequently assigned the highest posterior probability” (Hohwy, Roepstorff, and Friston (2008)) Perceptual level processing impacts on consciousness level processing and vice versa

Action-Oriented Predictive Processing Hierarchical predictive processing: include action (Friston, Daunizeau et al (2009), Friston (2010), Brown et al (2011)) Optimal feedback control theory: displays the motor control problem as mathematically equivalent to Bayesian inference (Todorov and Jordan (2002)) “Perceptual learning and inference is necessary to induce prior expectations about how the sensorium unfolds. Action is engaged to resample the world to fulfill these expectations. This places perception and action in intimate relation and accounts for both with the same principle” (Friston, Daunizeau, and Kiebel (2009) p. 12)

2. Representation, Inference, and the Continuity of Perception, Cognition, and Action 2.1 Explaining Away 2.2 Encoding, Inference, and the ‘Bayesian Brain’ 2.3 The Delicate Dance Between Top-Down and Bottom-Up

Explaining Away “High-level predictions explain away prediction error and tell the error units to ‘shut up’ [while] units encoding the causes of sensory input are selected by lateral interactions, with the error units, that mediate empirical priors.” (Friston (2005) p.829) Duplex architecture: it depicts the forward flow of information as solely conveying error, and the backwards flow as solely conveying predictions “Two functionally distinct subpopulations”: Encode the conditional expectations of perceptual causes and the prediction error respectively enabling an act between cancelling out and selective enhancement is only made possible (Friston (2005), p.829)

Encoding, Inference, and the ‘Bayesian Brain’ (1/3)

Encoding, Inference, and the ‘Bayesian Brain’ (2/3) Belletto, A View of Old Market in Schoessergasse (adapted from Daniel Dennett, Sweet Dream)

Encoding, Inference, and the ‘Bayesian Brain’ (3/3) For in many real-life cases, substantial context information is already in place when new information is encountered. An apt set of priors is thus often already active, poised to impact the processing of new sensory inputs without further delay. The brain, in ecologically normal circumstances, is not just suddenly ‘turned on’ and some random or unexpected input delivered for processing. So there is plenty of room for top-down influence to occur even before a stimulus is presented. Bayes’ Optimality: taking into account the uncertainty in our own sensory and motor signals, and adjusting the relative weight of different cues according to (often very subtle) contextual clues.

3. From Action-Oriented Predictive Processing to an Architecture of Mind 3.1 The Neural Evidence 3.2 Scope and Limits 3.3 Neats versus Scruffies (21st Century Replay). 3.4 Situated Agents

Challenges Evidential: What are the experimental and neuroanatomical implications, and to what extent are they borne out by current knowledge and investigations? Conceptual: Can we really explain so much about perception and action by direct appeal to a fundamental strategy of minimizing errors in the prediction of sensory input? Methodological: To what extent can these accounts hope to illuminate the full shape of the human cognitive architecture?

Situated Agents Synergies: Perception reduces surprisal by matching inputs with prior expectations. Action reduces surprisal by altering the world (including moving the body) so that inputs conform with expectations. Mobile robotics: show perception and behaviour productively interact via loops through action and the environment: create extra-neural opportunities for the minimization of prediction error “Behavioural feedback modifies stimulus sampling and so provides an additional extraneuronal path for the reduction of prediction errors”. (Verschure et al (2003) p.623)

Situated Agents Self-structuring of information flows: action-based structuring of sensory input (the linked unfolding across multiple sensory modalities that occurs when we see, touch, and hear an object actively manipulated) Promote learning and inference (Pfeifer, et al (2007), Clark (2008)) Robotic simulations in which the learning of complex co-ordination dynamics is achieved by maximizing the amount of predictive information present in sensorimotor loops. (Zahedi et al (in press)) "The architecture of the brain and the statistics of the environment are not fixed. Rather, brain-connectivity is subject to a broad spectrum of input-, experience-, and activity-dependent processes which shape and structure its patterning and strengths… These changes, in turn, result in altered interactions with the environment, exerting causal influences on what is experienced and sensed in the future” (Sporns (2007) p.179)

4. Content and Consciousness 4.1 Agency and Experience 4.2 Illuminating Experience: The Case of Delusions 4.3 Perception, Imagery, and the Senses 4.4 Sensing and World

5. Taking Stock 5.1 Comparison with Standard Computationalism 5.2 Conclusions: Towards A Grand Unified Theory of the Mind?

Summary Action-oriented (hierarchical) predictive processing model: bring cognition, perception, action, and attention together within a common framework. Suggesting probability density distributions induced by hierarchical generative models as basic means of representing the world, and prediction-error minimization as the driving force behind learning, action-selection, recognition, and inference. Specific phenomena: nonclassical receptive field effects, bi-stable perception, cue integration, and the pervasive context-sensitivity of neuronal response.

Summary Engagement with evolutionary, embodied, and situated approaches are needed. Prediction and prediction error minimization is powerful and illuminating. Unifying: Computational approaches (such as unsupervised and self-supervised forms of learning using recurrent neural network architectures), Probabilistic generative models for perception and action), Testable accounts of neural implementation  Addressing the levels of (in the vocabulary of Marr (1982)) the computation, the algorithm, and the implementation.