SepLL: Separating Latent Class Labels from Weak Supervision Noise

10/25/2022
by   Andreas Stephan, et al.
0

In the weakly supervised learning paradigm, labeling functions automatically assign heuristic, often noisy, labels to data samples. In this work, we provide a method for learning from weak labels by separating two types of complementary information associated with the labeling functions: information related to the target label and information specific to one labeling function only. Both types of information are reflected to different degrees by all labeled instances. In contrast to previous works that aimed at correcting or removing wrongly labeled instances, we learn a branched deep model that uses all data as-is, but splits the labeling function information in the latent space. Specifically, we propose the end-to-end model SepLL which extends a transformer classifier by introducing a latent space for labeling function specific and task-specific information. The learning signal is only given by the labeling functions matches, no pre-processing or label model is required for our method. Notably, the task prediction is made from the latent layer without any direct task signal. Experiments on Wrench text classification tasks show that our model is competitive with the state-of-the-art, and yields a new best average performance.

READ FULL TEXT
research
09/16/2021

KnowMAN: Weakly Supervised Multinomial Adversarial Networks

The absence of labeled data for training neural models is often addresse...
research
09/15/2020

Constrained Labeling for Weakly Supervised Learning

Curation of large fully supervised datasets has become one of the major ...
research
12/13/2022

Losses over Labels: Weakly Supervised Learning via Direct Loss Construction

Owing to the prohibitive costs of generating large amounts of labeled da...
research
03/21/2020

Crowdsourced Labeling for Worker-Task Specialization Block Model

We consider crowdsourced labeling under a worker-task specialization blo...
research
04/28/2022

WeaNF: Weak Supervision with Normalizing Flows

A popular approach to decrease the need for costly manual annotation of ...
research
05/08/2023

Q A Label Learning

Assigning labels to instances is crucial for supervised machine learning...
research
03/31/2020

Distance in Latent Space as Novelty Measure

Deep Learning performs well when training data densely covers the experi...

Please sign up or login with your details

Forgot password? Click here to reset