Continual Reinforcement Learning with associative memories
Thomas Schmied (Ph.D. Student)
The current generation of Deep Reinforcement learning systems is primarily designed to solve one particular task in a single stationary environment. However, the real world is non-stationary and dynamic by nature. For Reinforcement learning agents to be useful under these circumstances, they need to have the ability to efficiently learn a variety of tasks over an extended period of time in increasingly complex environments. To achieve this, Reinforcement learning agents must quickly adapt to changing environments, tasks, or distributions by leveraging the power of memory and context. In this project, we aim to develop novel, continual Reinforcement learning architectures that integrate dense associated memories, such as modern Hopfield networks, involved credit-assignment mechanisms, as well as recent advances in large-scale RL architectures via sequence modelling.
Primary Advisor: | Sepp Hochreiter (Johannes Kepler University Linz) |
Industry Advisor: | Razvan Pascanu (Google DeepMind) |
PhD Duration: | 01 February 2022 - 31 January 2025 |