A State Space Approach for Piecewise-Linear Recurrent Neural Networks for Reconstructing Nonlinear Dynamics from Neural Measurements

12/23/2016
by   Daniel Durstewitz, et al.
0

The computational properties of neural systems are often thought to be implemented in terms of their network dynamics. Hence, recovering the system dynamics from experimentally observed neuronal time series, like multiple single-unit (MSU) recordings or neuroimaging data, is an important step toward understanding its computations. Ideally, one would not only seek a state space representation of the dynamics, but would wish to have access to its governing equations for in-depth analysis. Recurrent neural networks (RNNs) are a computationally powerful and dynamically universal formal framework which has been extensively studied from both the computational and the dynamical systems perspective. Here we develop a semi-analytical maximum-likelihood estimation scheme for piecewise-linear RNNs (PLRNNs) within the statistical framework of state space models, which accounts for noise in both the underlying latent dynamics and the observation process. The Expectation-Maximization algorithm is used to infer the latent state distribution, through a global Laplace approximation, and the PLRNN parameters iteratively. After validating the procedure on toy examples, the approach is applied to MSU recordings from the rodent anterior cingulate cortex obtained during performance of a classical working memory task, delayed alternation. A model with 5 states turned out to be sufficient to capture the essential computational dynamics underlying task performance, including stimulus-selective delay activity. The estimated models were rarely multi-stable, but rather were tuned to exhibit slow dynamics in the vicinity of a bifurcation point. In summary, the present work advances a semi-analytical (thus reasonably fast) maximum-likelihood estimation framework for PLRNNs that may enable to recover the relevant dynamics underlying observed neuronal time series, and directly link them to computational properties.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2019

Identifying nonlinear dynamical systems via generative recurrent neural networks with applications to fMRI

A major tenet in theoretical neuroscience is that cognitive and behavior...
research
11/01/2021

Reverse engineering recurrent neural networks with Jacobian switching linear dynamical systems

Recurrent neural networks (RNNs) are powerful models for processing time...
research
06/29/2022

Hidden Parameter Recurrent State Space Models For Changing Dynamics Scenarios

Recurrent State-space models (RSSMs) are highly expressive models for le...
research
08/02/2021

Representation learning for neural population activity with Neural Data Transformers

Neural population activity is theorized to reflect an underlying dynamic...
research
01/08/2021

Slow manifolds in recurrent networks encode working memory efficiently and robustly

Working memory is a cognitive function involving the storage and manipul...
research
09/26/2022

Neural State-Space Modeling with Latent Causal-Effect Disentanglement

Despite substantial progress in deep learning approaches to time-series ...
research
06/07/2023

Generalized Teacher Forcing for Learning Chaotic Dynamics

Chaotic dynamical systems (DS) are ubiquitous in nature and society. Oft...

Please sign up or login with your details

Forgot password? Click here to reset