Download presentation
Presentation is loading. Please wait.
1
Latent Learning in Agents iCML 03 Robotics/Vision Workshop Rati Sharma
2
Problem statement Determine optimal paths in spatial navigation tasks. We use a deterministic grid environment as our world model. Various approaches have been used: ANN’s, Q-learning, Dyna
3
Latent Learning Tolman proposed the idea of Cognitive Maps based on experiments on Rats in a T-maze Latent Learning is described as learning that is not evident to the observer at the time it occurs but is apparent once a reinforcement is introduced.
4
Algorithms used: Q-Learning Q-Learning – Eventually converges to an optimal policy without ever having to learn and use an internal model of the environment Update rule Q(s,a) = r(s,a) + *max Q( (s,a),a’) r(s,a) is the reward function, (s,a) is the new state Disadvantage: Convergence to the optimal policy can be very slow
5
Model based learning- Dyna A form of planning is performed in addition to learning. Learning updates the appropriate value function estimates according to experience Planning updates the same value function estimates for simulated transitions chosen from the world model.
6
Problems considered Blocking problem Shortcut problem
7
Results
9
Conclusion Model based learning performs significantly better that Q-Learning. On the blocking and shortcut problems the agent demonstrates latent learning
10
Acknowledgements Prof. Littman Prof. Horatiu Voicu
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.