Fast Slow Learning: Incorporating Synthetic Gradients in Neural Memory Controllers

11/10/2020
by   Tharindu Fernando, et al.
0

Neural Memory Networks (NMNs) have received increased attention in recent years compared to deep architectures that use a constrained memory. Despite their new appeal, the success of NMNs hinges on the ability of the gradient-based optimiser to perform incremental training of the NMN controllers, determining how to leverage their high capacity for knowledge retrieval. This means that while excellent performance can be achieved when the training data is consistent and well distributed, rare data samples are hard to learn from as the controllers fail to incorporate them effectively during model training. Drawing inspiration from the human cognition process, in particular the utilisation of neuromodulators in the human brain, we propose to decouple the learning process of the NMN controllers to allow them to achieve flexible, rapid adaptation in the presence of new information. This trait is highly beneficial for meta-learning tasks where the memory controllers must quickly grasp abstract concepts in the target domain, and adapt stored knowledge. This allows the NMN controllers to quickly determine which memories are to be retained and which are to be erased, and swiftly adapt their strategy to the new task at hand. Through both quantitative and qualitative evaluations on multiple public benchmarks, including classification and regression tasks, we demonstrate the utility of the proposed approach. Our evaluations not only highlight the ability of the proposed NMN architecture to outperform the current state-of-the-art methods, but also provide insights on how the proposed augmentations help achieve such superior results. In addition, we demonstrate the practical implications of the proposed learning strategy, where the feedback path can be shared among multiple neural memory networks as a mechanism for knowledge sharing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/19/2023

Concept Discovery for Fast Adapatation

The advances in deep learning have enabled machine learning methods to o...
research
05/13/2019

Hierarchically Structured Meta-learning

In order to learn quickly with few samples, meta-learning utilizes prior...
research
03/02/2023

Model agnostic methods meta-learn despite misspecifications

Due to its empirical success on few shot classification and reinforcemen...
research
07/31/2023

MetaDiff: Meta-Learning with Conditional Diffusion for Few-Shot Learning

Equipping a deep model the abaility of few-shot learning, i.e., learning...
research
05/15/2019

LGM-Net: Learning to Generate Matching Networks for Few-Shot Learning

In this work, we propose a novel meta-learning approach for few-shot cla...
research
10/12/2019

Neural Memory Plasticity for Anomaly Detection

In the domain of machine learning, Neural Memory Networks (NMNs) have re...
research
05/27/2022

The Utility of Synthetic Reflexes and Haptic Feedback for Upper-Limb Prostheses in a Dexterous Task Without Direct Vision

Individuals who use myoelectric upper-limb prostheses often rely heavily...

Please sign up or login with your details

Forgot password? Click here to reset