Recurrent neural network models for working memory of continuous variables: activity manifolds, connectivity patterns, and dynamic codes

11/01/2021
by   Christopher J. Cueva, et al.
0

Many daily activities and psychophysical experiments involve keeping multiple items in working memory. When items take continuous values (e.g., orientation, contrast, length, loudness) they must be stored in a continuous structure of appropriate dimensions. We investigate how this structure is represented in neural circuits by training recurrent networks to report two previously shown stimulus orientations. We find the activity manifold for the two orientations resembles a Clifford torus. Although a Clifford and standard torus (the surface of a donut) are topologically equivalent, they have important functional differences. A Clifford torus treats the two orientations equally and keeps them in orthogonal subspaces, as demanded by the task, whereas a standard torus does not. We find and characterize the connectivity patterns that support the Clifford torus. Moreover, in addition to attractors that store information via persistent activity, our networks also use a dynamic code where units change their tuning to prevent new sensory input from overwriting the previously stored one. We argue that such dynamic codes are generally required whenever multiple inputs enter a memory system via shared connections. Finally, we apply our framework to a human psychophysics experiment in which subjects reported two remembered orientations. By varying the training conditions of the RNNs, we test and support the hypothesis that human behavior is a product of both neural noise and reliance on the more stable and behaviorally relevant memory of the ordinal relationship between the two orientations. This suggests that suitable inductive biases in RNNs are important for uncovering how the human brain implements working memory. Together, these results offer an understanding of the neural computations underlying a class of visual decoding tasks, bridging the scales from human behavior to synaptic connectivity.

READ FULL TEXT
research
08/27/2015

Continuous parameter working memory in a balanced chaotic neural network

It has been proposed that neural noise in the cortex arises from chaotic...
research
06/02/2018

A Novel Framework for Recurrent Neural Networks with Enhancing Information Processing and Transmission between Units

This paper proposes a novel framework for recurrent neural networks (RNN...
research
03/29/2022

Artificial Intelligence Software Structured to Simulate Human Working Memory, Mental Imagery, and Mental Continuity

This article presents an artificial intelligence (AI) architecture inten...
research
03/29/2022

A Computational Architecture for Machine Consciousness and Artificial Superintelligence: Updating Working Memory Iteratively

This theoretical article examines how to construct human-like working me...
research
08/24/2023

Persistent learning signals and working memory without continuous attractors

Neural dynamical systems with stable attractor structures, such as point...
research
12/10/2021

Where is Memory Information Stored in the Brain?

Within the scientific research community, memory information in the brai...
research
06/25/2020

Habit learning supported by efficiently controlled network dynamics in naive macaque monkeys

Primates display a marked ability to learn habits in uncertain and dynam...

Please sign up or login with your details

Forgot password? Click here to reset