Hierarchical Active Inference: A Theory of Motivated Control

Slides:



Advertisements
Similar presentations
How do emotion and motivation direct executive control? Luiz Pessoa Trends in Cognitive Sciences Volume 13, Issue 4, Pages (April 2009) DOI: /j.tics
Advertisements

Abstract We offer a formal treatment of choice behaviour based on the premise that agents minimise the expected free energy of future outcomes. Crucially,
Two views of brain function Marcus E. Raichle Trends in Cognitive Sciences Volume 14, Issue 4, Pages (April 2010) DOI: /j.tics
Nicolas Alzetta CoNGA: Cognition and Neuroscience Group of Antwerp
Soyoun Kim, Jaewon Hwang, Daeyeol Lee  Neuron 
Volume 53, Issue 1, Pages 9-16 (January 2007)
Xaq Pitkow, Dora E. Angelaki  Neuron 
Neurodegenerative Diseases Target Large-Scale Human Brain Networks
Effective Connectivity between Hippocampus and Ventromedial Prefrontal Cortex Controls Preferential Choices from Memory  Sebastian Gluth, Tobias Sommer,
A Sparse Object Coding Scheme in Area V4
Volume 80, Issue 4, Pages (November 2013)
Choosing Goals, Not Rules: Deciding among Rule-Based Action Plans
Volume 95, Issue 5, Pages e5 (August 2017)
Interoceptive inference, emotion, and the embodied self
Volume 59, Issue 6, Pages (September 2008)
Frontal Cortex and the Discovery of Abstract Action Rules
Alan N. Hampton, Ralph Adolphs, J. Michael Tyszka, John P. O'Doherty 
Decision Making: From Neuroscience to Psychiatry
Volume 92, Issue 5, Pages (December 2016)
Learning to Simulate Others' Decisions
Volume 96, Issue 2, Pages e4 (October 2017)
Keno Juechems, Jan Balaguer, Maria Ruz, Christopher Summerfield  Neuron 
Hierarchical Active Inference: A Theory of Motivated Control
Nils Kolling, Marco Wittmann, Matthew F.S. Rushworth  Neuron 
Volume 81, Issue 6, Pages (March 2014)
Arkady Konovalov, Ian Krajbich  Neuron 
Models of visual word recognition
Volume 59, Issue 6, Pages (September 2008)
Roman F. Loonis, Scott L. Brincat, Evan G. Antzoulatos, Earl K. Miller 
Feature- and Order-Based Timing Representations in the Frontal Cortex
Activity in Posterior Parietal Cortex Is Correlated with the Relative Subjective Desirability of Action  Michael C. Dorris, Paul W. Glimcher  Neuron 
Confidence as Bayesian Probability: From Neural Origins to Behavior
A Switching Observer for Human Perceptual Estimation
Decision Making: From Neuroscience to Psychiatry
Decoding Cognitive Processes from Neural Ensembles
Volume 65, Issue 6, Pages (March 2010)
Dynamic Coding for Cognitive Control in Prefrontal Cortex
Network Neuroscience Theory of Human Intelligence
Dharshan Kumaran, Hans Ludwig Melo, Emrah Duzel  Neuron 
Human Orbitofrontal Cortex Represents a Cognitive Map of State Space
Quantifying Biomolecule Diffusivity Using an Optimal Bayesian Method
Franco Pestilli, Marisa Carrasco, David J. Heeger, Justin L. Gardner 
Volume 96, Issue 2, Pages e4 (October 2017)
A Switching Observer for Human Perceptual Estimation
Caleb E. Strait, Tommy C. Blanchard, Benjamin Y. Hayden  Neuron 
Know Your Place: Neural Processing of Social Hierarchy in Humans
Neural Mechanisms Underlying Human Consensus Decision-Making
Volume 92, Issue 5, Pages (December 2016)
Neural Primacy of the Salience Processing System in Schizophrenia
The Social Dominance Paradox
Memory aging and brain maintenance
Franco Pestilli, Marisa Carrasco, David J. Heeger, Justin L. Gardner 
Volume 59, Issue 5, Pages (September 2008)
Timescales of Inference in Visual Adaptation
Janina Hesse, Susanne Schreiber  Current Biology 
Volume 76, Issue 4, Pages (November 2012)
Learning to Simulate Others' Decisions
Volume 27, Issue 3, Pages (September 2000)
Encoding of Stimulus Probability in Macaque Inferior Temporal Cortex
The Regulatory Role of the Human Mediodorsal Thalamus
Data simulation using the routine spm_MDP_VB.
Volume 62, Issue 4, Pages (May 2009)
Perceptual Classification in a Rapidly Changing Environment
Volume 23, Issue 21, Pages (November 2013)
Volume 13, Issue 9, Pages (September 2005)
Looking for cognition in the structure within the noise
An Introduction to Infinite HMMs for Single-Molecule Data Analysis
Volume 50, Issue 4, Pages (May 2006)
Medial Prefrontal Cortex Updates Its Status
Keno Juechems, Jan Balaguer, Maria Ruz, Christopher Summerfield  Neuron 
Presentation transcript:

Hierarchical Active Inference: A Theory of Motivated Control Giovanni Pezzulo, Francesco Rigoli, Karl J. Friston  Trends in Cognitive Sciences  Volume 22, Issue 4, Pages 294-306 (April 2018) DOI: 10.1016/j.tics.2018.01.009 Copyright © 2018 The Author(s) Terms and Conditions

Figure 1 An Example of Active Inference: The Waiting Game. The waiting game illustrates the importance of withholding prepotent responses [53]. At the beginning of the game, a low offer is available that can be replaced by a high offer or withdrawn. The player has prior preferences for ending up in the ‘accepted offer’ states, with a greater preference for the high offer. During the game, which lasts 16 trials, the player can thus either accept the low offer before it disappears or reject it and wait until it is converted to a high offer – which is risky, since the offer can be withdrawn. Active inference solves this problem by inferring both beliefs about hidden states of the game and the control policies (sequences of accept or reject/wait actions) that should be pursued [83–86]. The inference is based on a generative model (A), which includes five hidden states (circles) and their contingencies in terms of probabilistic transitions among hidden states, under different actions (edges). Accepting an offer moves the state to ‘accepted’ state (unless the offer has already been accepted or withdrawn). Rejecting a low offer (that has not been already been accepted or withdrawn) has three potential effects: it can be transformed into a high offer (with a low probability q), remain in play (with a high probability p) or be withdrawn (with a low probability r). The lower (B and C) panels show the results of simulating 16 trials, in which the low offer is converted to a high offer and accepted at the 10th choice point (B), or withdrawn on the 5th choice point (C). The top-left and top-right subpanels show expectations about which hidden state is occupied over time, and the expectation about accepting (green) or rejecting (blue) as time proceeds. The dotted lines correspond to beliefs about behaviour in the future formed during the game, while the solid lines show postdicted expectations at the end of the game. The bottom-left and bottom-right panels show the precision of policies and their deconvolution (to simulate dopaminergic responses) – which differ significantly when a preferred outcome can be attained (B) or not (C). See Box 1 for more details. Trends in Cognitive Sciences 2018 22, 294-306DOI: (10.1016/j.tics.2018.01.009) Copyright © 2018 The Author(s) Terms and Conditions

Figure 2 Simplified Belief Propagation Scheme for Deep Goal Hierarchies. The figure shows a hierarchical generative model that includes three levels, corresponding to three corticothalamic loops of increasing hierarchical depth [87], and the neuronal message passing that realizes a functional integration between dorsolateral and ventromedial structures (represented at a subcortical level for simplicity). The first and second equations mean that the higher levels of the control hierarchy [whose states are S(i+1)] prescribe the initial states S1(i) (via A) and the prior preferences over the evolution of their future sequences of states Sτ(i) (via C) of the lower levels. Crucially, the influence from higher- to lower-level state sequences is precision weighted, and the motivational hierarchy sets the precision γ(i) of top–down messages within the control hierarchy. This allows the motivational hierarchy to optimize the precision of prior preferences (or goals), by reflecting the incentives inferred at each level. At the lowest level, the states and trajectories specify set points for motor or autonomic control [88]. The third equation means that each level is equipped with probability transition matrices and policies or transition sequences (B), permitting to infer future states Sτ+1(i) based on the previous state at the same level Sτ(i) and the selected policy π(i). The latter equation shows that the probability of selecting a policy π(i) is treated as a Bayesian model selection problem, using a Softmax function of its expected evidence (free energy; G). The variables u, o and π denote motor actions, observations and policies or sequences of state transitions. Superscripts denote hierarchical levels. See [89] for more details. Trends in Cognitive Sciences 2018 22, 294-306DOI: (10.1016/j.tics.2018.01.009) Copyright © 2018 The Author(s) Terms and Conditions

Figure I Sensorimotor, Semantic and Episodic Contexts within Deep Goal Hierarchies. Amy, amygdala; dlPFC, dorsolateral prefrontal cortex; Hypo, hypothalamus; lPFC, lateral prefrontal cortex; PAG, periaqueductal gray; SMA, supplementary motor area; vlPFC, ventrolateral prefrontal cortex; vmPFC, ventromedial prefrontal cortex. Trends in Cognitive Sciences 2018 22, 294-306DOI: (10.1016/j.tics.2018.01.009) Copyright © 2018 The Author(s) Terms and Conditions