Prioritized Sweeping Neural DynaQ with Multiple Predecessors, and Hippocampal Replays

02/15/2018
by   Lise Aubin, et al.
0

During sleep and awake rest, the hippocampus replays sequences of place cells that have been activated during prior experiences. These have been interpreted as a memory consolidation process, but recent results suggest a possible interpretation in terms of reinforcement learning. The Dyna reinforcement learning algorithms use off-line replays to improve learning. Under limited replay budget, a prioritized sweeping approach, which requires a model of the transitions to the predecessors, can be used to improve performance. We investigate whether such algorithms can explain the experimentally observed replays. We propose a neural network version of prioritized sweeping Q-learning, for which we developed a growing multiple expert algorithm, able to cope with multiple predecessors. The resulting architecture is able to improve the learning of simulated agents confronted to a navigation task. We predict that, in animals, learning the world model should occur during rest periods, and that the corresponding replays should be shuffled.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/23/2021

A Robotic Model of Hippocampal Reverse Replay for Reinforcement Learning

Hippocampal reverse replay is thought to contribute to learning, and par...
research
05/30/2017

Experience Replay Using Transition Sequences

Experience replay is one of the most commonly used approaches to improve...
research
06/05/2022

Rapid Learning of Spatial Representations for Goal-Directed Navigation Based on a Novel Model of Hippocampal Place Fields

The discovery of place cells and other spatially modulated neurons in th...
research
06/12/2019

When to use parametric models in reinforcement learning?

We examine the question of when and how parametric models are most usefu...
research
09/18/2022

A Computational Model of Learning Flexible Navigation in a Maze by Layout-Conforming Replay of Place Cells

Recent experimental observations have shown that the reactivation of hip...
research
08/17/2023

Reinforced Self-Training (ReST) for Language Modeling

Reinforcement learning from human feedback (RLHF) can improve the qualit...
research
09/09/2021

Memory semantization through perturbed and adversarial dreaming

Classical theories of memory consolidation emphasize the importance of r...

Please sign up or login with your details

Forgot password? Click here to reset