Foreshadowing the Benefits of Incidental Supervision

06/09/2020
by   Hangfeng He, et al.
19

Learning theory mostly addresses the standard learning paradigm, assuming the availability of complete and correct supervision signals for large amounts of data. However, in practice, machine learning researchers and practitioners acquire and make use of a range of incidental supervision signals that only have statistical associations with the gold supervision. This paper addresses the question: Can one quantify models' performance when learning with such supervision signals, without going through an exhaustive experimentation process with various supervision signals and learning protocols? To quantify the benefits of various incidental supervision signals, we propose a unified PAC-Bayesian Informativeness measure (PABI), characterizing the reduction in uncertainty that incidental supervision signals provide. We then demonstrate PABI's use in quantifying various types of incidental signals such as partial labels, noisy labels, constraints, cross-domain signals, and some combinations of these. Experiments on named entity recognition and question answering show that PABI correlates well with learning performance, providing a promising way to determine, ahead of learning, which supervision signals would be beneficial.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2022

Data Consistency for Weakly Supervised Learning

In many applications, training machine learning models involves using la...
research
06/15/2020

Learnability with Indirect Supervision Signals

Learning from indirect supervision signals is important in real-world AI...
research
09/10/2018

Learning Named Entity Tagger using Domain-Specific Dictionary

Recent advances in deep neural models allow us to build reliable named e...
research
10/09/2022

Contrastive Representation Learning for Conversational Question Answering over Knowledge Graphs

This paper addresses the task of conversational question answering (Conv...
research
06/01/2021

Reinforced Iterative Knowledge Distillation for Cross-Lingual Named Entity Recognition

Named entity recognition (NER) is a fundamental component in many applic...
research
01/11/2021

Improving Multi-hop Knowledge Base Question Answering by Learning Intermediate Supervision Signals

Multi-hop Knowledge Base Question Answering (KBQA) aims to find the answ...

Please sign up or login with your details

Forgot password? Click here to reset