Time & Date 31.10.2019
5-7 ct
Room 47.0.501 (Teaching block WWP)
Universität West
Albert-Einstein-Allee 47
89081 Ulm
Links:
Cognitive Systems and Human-Computer Interaction
Cognitive Systems M.Sc.
Abstract. Can a slow, incremental learning rule give rise to a sophisticated reinforcement learning algorithm? I will show how a deep learning system with memory, when trained on a distribution of tasks, learns a new learning algorithm which turns out to be Bayes-optimal. In particular, we will see how these algorithms are implemented in the memory, and relate it to Bayesian reinforcement learning.
Bio. Pedro A. Ortega is a Research Scientist at DeepMind. His work includes the application of information-theoretic and statistical mechanical ideas to sequential decision-making, which has led to contributions in novel bounded rationality models and recasting adaptive control as a causal inference problem. He obtained his PhD in Engineering from the University of Cambridge (Prof. Zoubin Ghahramani), and he has been a post-doctoral fellow at the Department of Engineering in Cambridge (Prof. Simon Godsill), at the Max Planck Institute for Biological Cybernetics/Intelligent Systems (Daniel A. Braun), at the Hebrew University in Jerusalem (Prof. Naftali Tishby), and at the University of Pennsylvania (Prof. Daniel D. Lee and Prof. Alan A. Stocker).
Time & Date 31.10.2019
5-7 ct
Room 47.0.501 (Teaching block WWP)
Universität West
Albert-Einstein-Allee 47
89081 Ulm
Links:
Cognitive Systems and Human-Computer Interaction
Cognitive Systems M.Sc.