Gumbel-Softmax Selective Networks

11/19/2022
by   Mahmoud Salem, et al.
0

ML models often operate within the context of a larger system that can adapt its response when the ML model is uncertain, such as falling back on safe defaults or a human in the loop. This commonly encountered operational context calls for principled techniques for training ML models with the option to abstain from predicting when uncertain. Selective neural networks are trained with an integrated option to abstain, allowing them to learn to recognize and optimize for the subset of the data distribution for which confident predictions can be made. However, optimizing selective networks is challenging due to the non-differentiability of the binary selection function (the discrete decision of whether to predict or abstain). This paper presents a general method for training selective networks that leverages the Gumbel-softmax reparameterization trick to enable selection within an end-to-end differentiable training framework. Experiments on public datasets demonstrate the potential of Gumbel-softmax selective networks for selective regression and classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2019

SelectiveNet: A Deep Neural Network with an Integrated Reject Option

We consider the problem of selective prediction (also known as reject op...
research
05/23/2017

Selective Classification for Deep Neural Networks

Selective classification techniques (also known as reject option) have n...
research
10/29/2018

Learning to Screen for Fast Softmax Inference on Large Vocabulary Neural Networks

Neural language models have been widely used in various NLP tasks, inclu...
research
12/23/2022

Neural Networks beyond explainability: Selective inference for sequence motifs

Over the past decade, neural networks have been successful at making pre...
research
05/27/2023

On the special role of class-selective neurons in early training

It is commonly observed that deep networks trained for classification ex...
research
11/16/2019

Selective sampling for accelerating training of deep neural networks

We present a selective sampling method designed to accelerate the traini...
research
07/02/2018

Learning under selective labels in the presence of expert consistency

We explore the problem of learning under selective labels in the context...

Please sign up or login with your details

Forgot password? Click here to reset