B-SMALL: A Bayesian Neural Network approach to Sparse Model-Agnostic Meta-Learning

01/01/2021
by   Anish Madan, et al.
0

There is a growing interest in the learning-to-learn paradigm, also known as meta-learning, where models infer on new tasks using a few training examples. Recently, meta-learning based methods have been widely used in few-shot classification, regression, reinforcement learning, and domain adaptation. The model-agnostic meta-learning (MAML) algorithm is a well-known algorithm that obtains model parameter initialization at meta-training phase. In the meta-test phase, this initialization is rapidly adapted to new tasks by using gradient descent. However, meta-learning models are prone to overfitting since there are insufficient training tasks resulting in over-parameterized models with poor generalization performance for unseen tasks. In this paper, we propose a Bayesian neural network based MAML algorithm, which we refer to as the B-SMALL algorithm. The proposed framework incorporates a sparse variational loss term alongside the loss function of MAML, which uses a sparsifying approximated KL divergence as a regularizer. We demonstrate the performance of B-MAML using classification and regression tasks, and highlight that training a sparsifying BNN using MAML indeed improves the parameter footprint of the model while performing at par or even outperforming the MAML approach. We also illustrate applicability of our approach in distributed sensor networks, where sparsity and meta-learning can be beneficial.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/13/2020

Regularizing Meta-Learning via Gradient Dropout

With the growing attention on learning-to-learn new tasks using only a f...
research
10/16/2021

Meta-Learning with Adjoint Methods

Model Agnostic Meta-Learning (MAML) is widely used to find a good initia...
research
07/07/2020

Meta-Learning with Network Pruning

Meta-learning is a powerful paradigm for few-shot learning. Although wit...
research
12/14/2018

Online gradient-based mixtures for transfer modulation in meta-learning

Learning-to-learn or meta-learning leverages data-driven inductive bias ...
research
03/15/2021

Robust MAML: Prioritization task buffer with adaptive learning process for model-agnostic meta-learning

Model agnostic meta-learning (MAML) is a popular state-of-the-art meta-l...
research
12/10/2020

Performance-Weighed Policy Sampling for Meta-Reinforcement Learning

This paper discusses an Enhanced Model-Agnostic Meta-Learning (E-MAML) a...
research
02/25/2016

Meta-learning within Projective Simulation

Learning models of artificial intelligence can nowadays perform very wel...

Please sign up or login with your details

Forgot password? Click here to reset