Less is More – Towards parsimonious multi-task models using structured sparsity

08/23/2023
by   Richa Upadhyay, et al.
0

Group sparsity in Machine Learning (ML) encourages simpler, more interpretable models with fewer active parameter groups. This work aims to incorporate structured group sparsity into the shared parameters of a Multi-Task Learning (MTL) framework, to develop parsimonious models that can effectively address multiple tasks with fewer parameters while maintaining comparable or superior performance to a dense model. Sparsifying the model during training helps decrease the model's memory footprint, computation requirements, and prediction time during inference. We use channel-wise l1/l2 group sparsity in the shared layers of the Convolutional Neural Network (CNN). This approach not only facilitates the elimination of extraneous groups (channels) but also imposes a penalty on the weights, thereby enhancing the learning of all tasks. We compare the outcomes of single-task and multi-task experiments under group sparsity on two publicly available MTL datasets, NYU-v2 and CelebAMask-HQ. We also investigate how changing the sparsification degree impacts both the performance of the model and the sparsity of groups.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/24/2020

Latent Group Structured Multi-task Learning

In multi-task learning (MTL), we improve the performance of key machine ...
research
06/27/2012

Learning Task Grouping and Overlap in Multi-task Learning

In the paradigm of multi-task learning, mul- tiple related prediction ta...
research
03/10/2015

Learning the Structure for Structured Sparsity

Structured sparsity has recently emerged in statistics, machine learning...
research
03/10/2022

A Tree-Structured Multi-Task Model Recommender

Tree-structured multi-task architectures have been employed to jointly t...
research
02/09/2020

Multi-Task Learning by a Top-Down Control Network

A general problem that received considerable recent attention is how to ...
research
08/15/2012

Efficient Algorithm for Extremely Large Multi-task Regression with Massive Structured Sparsity

We develop a highly scalable optimization method called "hierarchical gr...
research
05/21/2022

Multi-task Learning for Gaussian Graphical Regressions with High Dimensional Covariates

Gaussian graphical regression is a powerful means that regresses the pre...

Please sign up or login with your details

Forgot password? Click here to reset