Neuromorphic Architecture Optimization for Task-Specific Dynamic Learning

06/04/2019
by   Sandeep Madireddy, et al.
0

The ability to learn and adapt in real time is a central feature of biological systems. Neuromorphic architectures demonstrating such versatility can greatly enhance our ability to efficiently process information at the edge. A key challenge, however, is to understand which learning rules are best suited for specific tasks and how the relevant hyperparameters can be fine-tuned. In this work, we introduce a conceptual framework in which the learning process is integrated into the network itself. This allows us to cast meta-learning as a mathematical optimization problem. We employ DeepHyper, a scalable, asynchronous model-based search, to simultaneously optimize the choice of meta-learning rules and their hyperparameters. We demonstrate our approach with two different datasets, MNIST and FashionMNIST, using a network architecture inspired by the learning center of the insect brain. Our results show that optimal learning rules can be dataset-dependent even within similar tasks. This dependency demonstrates the importance of introducing versatility and flexibility in the learning algorithms. It also illuminates experimental findings in insect neuroscience that have shown a heterogeneity of learning rules within the insect mushroom body.

READ FULL TEXT
research
03/15/2019

Neuromorphic Hardware learns to learn

Hyperparameters and learning algorithms for neuromorphic hardware are us...
research
10/19/2020

Meta-learning the Learning Trends Shared Across Tasks

Meta-learning stands for 'learning to learn' such that generalization to...
research
11/13/2020

Testing the Genomic Bottleneck Hypothesis in Hebbian Meta-Learning

Recent work has shown promising results using Hebbian meta-learning to s...
research
02/07/2020

Meta-learning framework with applications to zero-shot time-series forecasting

Can meta-learning discover generic ways of processing time-series (TS) f...
research
01/26/2022

Meta-learning Spiking Neural Networks with Surrogate Gradient Descent

Adaptive "life-long" learning at the edge and during online task perform...
research
07/16/2020

Multilayer Neuromodulated Architectures for Memory-Constrained Online Continual Learning

We focus on the problem of how to achieve online continual learning unde...

Please sign up or login with your details

Forgot password? Click here to reset