Meta-Learning via Feature-Label Memory Network

10/19/2017
by   Dawit Mureja, et al.
0

Deep learning typically requires training a very capable architecture using large datasets. However, many important learning problems demand an ability to draw valid inferences from small size datasets, and such problems pose a particular challenge for deep learning. In this regard, various researches on "meta-learning" are being actively conducted. Recent work has suggested a Memory Augmented Neural Network (MANN) for meta-learning. MANN is an implementation of a Neural Turing Machine (NTM) with the ability to rapidly assimilate new data in its memory, and use this data to make accurate predictions. In models such as MANN, the input data samples and their appropriate labels from previous step are bound together in the same memory locations. This often leads to memory interference when performing a task as these models have to retrieve a feature of an input from a certain memory location and read only the label information bound to that location. In this paper, we tried to address this issue by presenting a more robust MANN. We revisited the idea of meta-learning and proposed a new memory augmented neural network by explicitly splitting the external memory into feature and label memories. The feature memory is used to store the features of input data samples and the label memory stores their labels. Hence, when predicting the label of a given input, our model uses its feature memory unit as a reference to extract the stored feature of the input, and based on that feature, it retrieves the label information of the input from the label memory unit. In order for the network to function in this framework, a new memory-writingmodule to encode label information into the label memory in accordance with the meta-learning task structure is designed. Here, we demonstrate that our model outperforms MANN by a large margin in supervised one-shot classification tasks using Omniglot and MNIST datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/26/2021

MCML: A Novel Memory-based Contrastive Meta-Learning Method for Few Shot Slot Tagging

Meta-learning is widely used for few-shot slot tagging in the task of fe...
research
06/08/2023

EMO: Episodic Memory Optimization for Few-Shot Meta-Learning

Few-shot meta-learning presents a challenge for gradient descent optimiz...
research
10/07/2019

Meta-Learning Deep Energy-Based Memory Models

We study the problem of learning associative memory – a system which is ...
research
02/10/2019

Task2Vec: Task Embedding for Meta-Learning

We introduce a method to provide vectorial representations of visual cla...
research
07/05/2017

Labeled Memory Networks for Online Model Adaptation

Augmenting a neural network with memory that can grow without growing th...
research
09/07/2020

Information Theoretic Meta Learning with Gaussian Processes

We formulate meta learning using information theoretic concepts such as ...
research
12/28/2022

Learning to Detect Noisy Labels Using Model-Based Features

Label noise is ubiquitous in various machine learning scenarios such as ...

Please sign up or login with your details

Forgot password? Click here to reset