Contextual Classification Using Self-Supervised Auxiliary Models for Deep Neural Networks

01/07/2021
by   Sebastian Palacio, et al.
0

Classification problems solved with deep neural networks (DNNs) typically rely on a closed world paradigm, and optimize over a single objective (e.g., minimization of the cross-entropy loss). This setup dismisses all kinds of supporting signals that can be used to reinforce the existence or absence of a particular pattern. The increasing need for models that are interpretable by design makes the inclusion of said contextual signals a crucial necessity. To this end, we introduce the notion of Self-Supervised Autogenous Learning (SSAL) models. A SSAL objective is realized through one or more additional targets that are derived from the original supervised classification task, following architectural principles found in multi-task learning. SSAL branches impose low-level priors into the optimization process (e.g., grouping). The ability of using SSAL branches during inference, allow models to converge faster, focusing on a richer set of class-relevant features. We show that SSAL models consistently outperform the state-of-the-art while also providing structured predictions that are more interpretable.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/13/2023

Semi-supervised learning made simple with self-supervised clustering

Self-supervised learning models have been shown to learn rich visual rep...
research
10/22/2020

Perceptual Loss based Speech Denoising with an ensemble of Audio Pattern Recognition and Self-Supervised Models

Deep learning based speech denoising still suffers from the challenge of...
research
11/14/2020

Graph-Based Neural Network Models with Multiple Self-Supervised Auxiliary Tasks

Self-supervised learning is currently gaining a lot of attention, as it ...
research
07/13/2022

Task Agnostic Representation Consolidation: a Self-supervised based Continual Learning Approach

Continual learning (CL) over non-stationary data streams remains one of ...
research
06/11/2019

Self-Supervised Learning for Contextualized Extractive Summarization

Existing models for extractive summarization are usually trained from sc...

Please sign up or login with your details

Forgot password? Click here to reset