Abstract We offer a formal treatment of choice behaviour based on the premise that agents minimise the expected free energy of future outcomes. Crucially,

Slides:



Advertisements
Similar presentations
Bayesian inference Lee Harrison York Neuroimaging Centre 01 / 05 / 2009.
Advertisements

Bayesian inference Jean Daunizeau Wellcome Trust Centre for Neuroimaging 16 / 05 / 2008.
Bayesian models for fMRI data
Bayesian Inference Chris Mathys Wellcome Trust Centre for Neuroimaging UCL SPM Course London, May 12, 2014 Thanks to Jean Daunizeau and Jérémie Mattout.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
MEG/EEG Inverse problem and solutions In a Bayesian Framework EEG/MEG SPM course, Bruxelles, 2011 Jérémie Mattout Lyon Neuroscience Research Centre ? ?
Chapter 4: Linear Models for Classification
1 Slides for the book: Probabilistic Robotics Authors: Sebastian Thrun Wolfram Burgard Dieter Fox Publisher: MIT Press, Web site for the book & more.
The free-energy principle: a rough guide to the brain? Karl Friston
Yiannis Demiris and Anthony Dearden By James Gilbert.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Bayesian network inference
CS 589 Information Risk Management 6 February 2007.
Dutch books and epistemic events Jan-Willem Romeijn Psychological Methods University of Amsterdam ILLC 2005 Interfacing Probabilistic and Epistemic Update.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
. PGM: Tirgul 10 Parameter Learning and Priors. 2 Why learning? Knowledge acquisition bottleneck u Knowledge acquisition is an expensive process u Often.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Reward processing (1) There exists plenty of evidence that midbrain dopamine systems encode errors in reward predictions (Schultz, Neuron, 2002) Changes.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to understand.
Free energy and active inference Karl Friston Abstract How much about our interaction with – and experience of – our world can be deduced from basic principles?
Computational and Physiological Models Part 1
Abstract We start with a statistical formulation of Helmholtz’s ideas about neural energy to furnish a model of perceptual inference and learning that.
The free-energy principle: a rough guide to the brain? K Friston Summarized by Joon Shik Kim (Thu) Computational Models of Intelligence.
Statistical Decision Theory
ABSTRACT: My treatment of critical gaps in models of probabilistic inference will focus on the potential of unified theories to “close the gaps” between.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
Probabilistic Graphical Models
Mean Field Variational Bayesian Data Assimilation EGU 2012, Vienna Michail Vrettas 1, Dan Cornford 1, Manfred Opper 2 1 NCRG, Computer Science, Aston University,
Abstract This talk summarizes recent attempts to integrate action and perception within a single optimization framework. We start with a statistical formulation.
Bayesian Inference and Posterior Probability Maps Guillaume Flandin Wellcome Department of Imaging Neuroscience, University College London, UK SPM Course,
Abstract This talk summarizes our recent attempts to integrate action and perception within a single optimization framework. We start with a statistical.
Abstract: How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Abstract Predictive coding models and the free-energy principle, suggests that cortical activity in sensory brain areas reflects the precision of prediction.
Abstract: This overview of the free energy principle offers an account of embodied exchange with the world that associates conscious operations with actively.
Abstract This presentation questions the need for reinforcement learning and related paradigms from machine-learning, when trying to optimise the behavior.
Abstract How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts.
Free energy and active inference
Reasons to be careful about reward A flow (policy) cannot be specified with a scalar function of states: the fundamental theorem of vector calculus – aka.
Bayesian Approach For Clinical Trials Mark Chang, Ph.D. Executive Director Biostatistics and Data management AMAG Pharmaceuticals Inc.
Dropout as a Bayesian Approximation
Free-energy and active inference
Abstract We suggested recently that attention can be understood as inferring the level of uncertainty or precision during hierarchical perception. In.
Abstract This presentation will look at action, perception and cognition as emergent phenomena under a unifying perspective: This Helmholtzian perspective.
Workshop on Mathematical Models of Cognitive Architectures December 5-9, 2011 CIRM, Marseille Workshop on Mathematical Models of Cognitive Architectures.
Abstract In this presentation, I will rehearse the free-energy formulation of action and perception, with a special focus on the representation of uncertainty:
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Abstract In this presentation, I will rehearse the free-energy formulation of action and perception, with a special focus on the representation of uncertainty:
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
WP6 - D6.1 Design of integrated models ISTC-CNR September, 26/27, 2005 ISTC-CNR September, 26/27, 2005.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
Bayesian inference Lee Harrison York Neuroimaging Centre 23 / 10 / 2009.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Abstract How much about our interactions with – and experience of – our world can be deduced from basic principles? This talk reviews recent attempts to.
Probabilistic Robotics Probability Theory Basics Error Propagation Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Tutorial Session: The Bayesian brain, surprise and free-energy Value-learning and perceptual learning have been an important focus over the past decade,
Variational filtering in generated coordinates of motion
Computational models for imaging analyses
Wellcome Trust Centre for Neuroimaging University College London
Henrik Singmann Karl Christoph Klauer Sieghard Beller
The free-energy principle: a rough guide to the brain? K Friston
Hierarchical Active Inference: A Theory of Motivated Control
A schematic overview of the analysis stream underlying the treatment of computational psychiatry in this article. A schematic overview of the analysis.
Hierarchical Active Inference: A Theory of Motivated Control
Data simulation using the routine spm_MDP_VB.
Presentation transcript:

Abstract We offer a formal treatment of choice behaviour based on the premise that agents minimise the expected free energy of future outcomes. Crucially, the negative free energy or quality of a policy can be decomposed into extrinsic and epistemic (intrinsic) value. Minimising expected free energy is therefore equivalent to maximising extrinsic value or expected utility (defined in terms of prior preferences or goals), while maximising information gain or intrinsic value; i.e., reducing uncertainty about the causes of valuable outcomes. The resulting scheme resolves the exploration-exploitation dilemma: epistemic value is maximised until there is no further information gain, after which exploitation is assured through maximisation of extrinsic value. This is formally consistent with the Infomax principle, generalising formulations of active vision based upon salience (Bayesian surprise) and optimal decisions based on expected utility and risk sensitive (KL) control. Furthermore, as with previous active inference formulations of discrete (Markovian) problems; ad hoc softmax parameters become the expected (Bayes-optimal) precision of beliefs about – or confidence in – policies. We focus on the basic theory – illustrating the minimisation of expected free energy using simulations. A key aspect of this minimisation is the similarity of precision updates and dopaminergic discharges observed in conditioning paradigms. Active inference and epistemic value Karl Friston, Francesco Rigoli, Dimitri Ognibene, Christoph Mathys, Thomas FitzGerald and Giovanni Pezzulo

Premise  All agents minimize free energy (under a generative model)  All agents possess prior beliefs (preferences)  Free energy is minimized when priors are (actively) realized  All agents believe they will minimize (expected) free energy

Set-up and definitions: active inference Free energy if action perception world agent Generative process Generative model Approximate posterior Perception-action cycle

Hidden states Control states ? ? Reject or stay Low offer High offer ? ? Accept or shift Low offer High offer An example:

if Full priors – control states Empirical priors – hidden states Likelihood The (normal form) generative model Hidden states Action Control states

KL or risk-sensitive control In the absence of ambiguity: Expected utility theory In the absence of posterior uncertainty or risk: Priors over policies Bayesian surprise and Infomax In the absence of prior beliefs about outcomes: Prior beliefs about policies Expected free energy Extrinsic value Epistemic value Predicted divergence Predicted ambiguity Predicted divergence Extrinsic value Bayesian surprise Predicted mutual information

The quality of a policy corresponds to (negative) expected free energy Generative model of future statesFuture generative model of states Prior preferences (goals) over future outcomes

The mean field partition And variational updates Minimising free energy

midbrain motor Cortex occipital Cortex striatum Variational updates Perception Action selection Precision Functional anatomy Forward sweeps over future states if prefrontal Cortex hippocampus Predicted divergence Predicted ambiguity

The (T-maze) problem Y Y or

YY Generative model Hidden states Control states Observations Extrinsic value Epistemic value Prior beliefs about control Posterior beliefs about states YYYY location stimulus CS CS US NS YYYY location context YYYY location

Comparing different schemes Performance Prior preference success rate (%) FE KL EU DA Sensitive to risk or ambiguity

Simulating conditioned responses (in terms of precision updates)

Expected value Expected precision Expected precision and value Changes in expected precision reflect changes in expected value: c.f., dopamine and reward prediction error

Simulating conditioned responses

Learning as inference YY Hidden states YYYY location context Control states YYYY YYYY maze Hierarchical augmentation of state space Bayesian belief updating between trials: of conserved (maze) states

Learning as inference performance uncertainty

Summary  Optimal behaviour can be cast as a pure inference problem, in which valuable outcomes are defined in terms of prior beliefs about future states.  Exact Bayesian inference (perfect rationality) cannot be realised physically, which means that optimal behaviour rests on approximate Bayesian inference (bounded rationality).  Variational free energy provides a bound on Bayesian model evidence that is optimised by bounded rational behaviour.  Bounded rational behaviour requires (approximate Bayesian) inference on both hidden states of the world and (future) control states. This mandates beliefs about action (control) that are distinct from action per se – beliefs that entail a precision.  These beliefs can be cast in terms of minimising the expected free energy, given current beliefs about the state of the world and future choices.  The ensuing quality of a policy entails epistemic value and expected utility that account for exploratory and exploitative behaviour respectively.  Variational Bayes provides a formal account of how posterior expectations about hidden states of the world, policies and precision depend upon each other; and may provide a metaphor for message passing in the brain.  Beliefs about choices depend upon expected precision while beliefs about precision depend upon the expected quality of choices.  Variational Bayes induces distinct probabilistic representations (functional segregation) of hidden states, control states and precision – and highlights the role of reciprocal message passing. This may be particularly important for expected precision that is required for optimal inference about hidden states (perception) and control states (action selection).  The dynamics of precision updates and their computational architecture are consistent with the physiology and anatomy of the dopaminergic system.