Learning as Reinforcement: Applying Principles of Neuroscience for More General Reinforcement Learning Agents

04/20/2020
by   Eric Zelikman, et al.
1

A significant challenge in developing AI that can generalize well is designing agents that learn about their world without being told what to learn, and apply that learning to challenges with sparse rewards. Moreover, most traditional reinforcement learning approaches explicitly separate learning and decision making in a way that does not correspond to biological learning. We implement an architecture founded in principles of experimental neuroscience, by combining computationally efficient abstractions of biological algorithms. Our approach is inspired by research on spike-timing dependent plasticity, the transition between short and long term memory, and the role of various neurotransmitters in rewarding curiosity. The Neurons-in-a-Box architecture can learn in a wholly generalizable manner, and demonstrates an efficient way to build and apply representations without explicitly optimizing over a set of criteria or actions. We find it performs well in many environments including OpenAI Gym's Mountain Car, which has no reward besides touching a hard-to-reach flag on a hill, Inverted Pendulum, where it learns simple strategies to improve the time it holds a pendulum up, a video stream, where it spontaneously learns to distinguish an open and closed hand, as well as other environments like Google Chrome's Dinosaur Game.

READ FULL TEXT

page 5

page 6

research
01/27/2020

Reinforcement Learning-based Autoscaling of Workflows in the Cloud: A Survey

Reinforcement Learning (RL) has demonstrated a great potential for autom...
research
10/24/2022

Evaluating Long-Term Memory in 3D Mazes

Intelligent agents need to remember salient information to reason in par...
research
09/21/2021

Learning offline: memory replay in biological and artificial reinforcement learning

Learning to act in an environment to maximise rewards is among the brain...
research
06/08/2021

There Is No Turning Back: A Self-Supervised Approach for Reversibility-Aware Reinforcement Learning

We propose to learn to distinguish reversible from irreversible actions ...
research
06/26/2020

What can I do here? A Theory of Affordances in Reinforcement Learning

Reinforcement learning algorithms usually assume that all actions are al...
research
06/21/2019

Split Q Learning: Reinforcement Learning with Two-Stream Rewards

Drawing an inspiration from behavioral studies of human decision making,...
research
04/14/2023

Robust Decision-Making in Spatial Learning: A Comparative Study of Successor Features and Predecessor Features Algorithms

Predictive map theory, one of the theories explaining spatial learning i...

Please sign up or login with your details

Forgot password? Click here to reset