Estimating scale-invariant future in continuous time

02/18/2018
by   Zoran Tiganj, et al.
0

Natural learners must compute an estimate of future outcomes that follow from a stimulus in continuous time. Critically, the learner cannot in general know a priori the relevant time scale over which meaningful relationships will be observed. Widely used reinforcement learning algorithms discretize continuous time and use the Bellman equation to estimate exponentially-discounted future reward. However, exponential discounting introduces a time scale to the computation of value. Scaling is a serious problem in continuous time: efficient learning with scaled algorithms requires prior knowledge of the relevant scale. That is, with scaled algorithms one must know at least part of the solution to a problem prior to attempting a solution. We present a computational mechanism, developed based on work in psychology and neuroscience, for computing a scale-invariant timeline of future events. This mechanism efficiently computes a model for future time on a logarithmically-compressed scale, and can be used to generate a scale-invariant power-law-discounted estimate of expected future reward. Moreover, the representation of future time retains information about what will happen when, enabling flexible decision making based on future events. The entire timeline can be constructed in a single parallel operation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/02/2020

POMDPs in Continuous Time and Discrete Spaces

Many processes, such as discrete event systems in engineering or populat...
research
01/26/2021

Predicting the future with a scale-invariant temporal memory for the past

In recent years it has become clear that the brain maintains a temporal ...
research
08/03/2022

Quantum Analysis of Continuous Time Stochastic Process

The continuous time stochastic process is a mainstream mathematical inst...
research
03/31/2023

An Efficient Off-Policy Reinforcement Learning Algorithm for the Continuous-Time LQR Problem

In this paper, an off-policy reinforcement learning algorithm is designe...
research
09/06/2023

Near-continuous time Reinforcement Learning for continuous state-action spaces

We consider the Reinforcement Learning problem of controlling an unknown...
research
11/22/2012

Optimally fuzzy temporal memory

Any learner with the ability to predict the future of a structured time-...
research
10/14/2022

Model-Free Characterizations of the Hamilton-Jacobi-Bellman Equation and Convex Q-Learning in Continuous Time

Convex Q-learning is a recent approach to reinforcement learning, motiva...

Please sign up or login with your details

Forgot password? Click here to reset