Learning Neural Sequence-to-Sequence Models from Weak Feedback with Bipolar Ramp Loss

07/06/2019
by   Laura Jehl, et al.
0

In many machine learning scenarios, supervision by gold labels is not available and consequently neural models cannot be trained directly by maximum likelihood estimation (MLE). In a weak supervision scenario, metric-augmented objectives can be employed to assign feedback to model outputs, which can be used to extract a supervision signal for training. We present several objectives for two separate weakly supervised tasks, machine translation and semantic parsing. We show that objectives should actively discourage negative outputs in addition to promoting a surrogate gold structure. This notion of bipolarity is naturally present in ramp loss objectives, which we adapt to neural models. We show that bipolar ramp loss objectives outperform other non-bipolar ramp loss objectives and minimum risk training (MRT) on both weakly supervised tasks, as well as on a supervised machine translation task. Additionally, we introduce a novel token-level ramp loss objective, which is able to outperform even the best sequence-level ramp loss on both weakly supervised tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/22/2015

Bandit Label Inference for Weakly Supervised Learning

The scarcity of data annotated at the desired level of granularity is a ...
research
04/23/2021

Knodle: Modular Weakly Supervised Learning with PyTorch

Methods for improving the training and prediction quality of weakly supe...
research
06/04/2020

MLE-guided parameter search for task loss minimization in neural sequence modeling

Neural autoregressive sequence models are used to generate sequences in ...
research
06/28/2017

(Machine) Learning to Do More with Less

Determining the best method for training a machine learning algorithm is...
research
06/20/2023

Open Problem: Learning with Variational Objectives on Measures

The theory of statistical learning has focused on variational objectives...
research
10/31/2018

Weakly Supervised Grammatical Error Correction using Iterative Decoding

We describe an approach to Grammatical Error Correction (GEC) that is ef...
research
05/19/2022

LAGr: Label Aligned Graphs for Better Systematic Generalization in Semantic Parsing

Semantic parsing is the task of producing structured meaning representat...

Please sign up or login with your details

Forgot password? Click here to reset