Avoiding Your Teacher's Mistakes: Training Neural Networks with Controlled Weak Supervision

11/01/2017
by   Mostafa Dehghani, et al.
0

Training deep neural networks requires massive amounts of training data, but for many tasks only limited labeled data is available. This makes weak supervision attractive, using weak or noisy signals like the output of heuristic methods or user click-through data for training. In a semi-supervised setting, we can use a large set of data with weak labels to pretrain a neural network and then fine-tune the parameters with a small amount of data with true labels. This feels intuitively sub-optimal as these two independent stages leave the model unaware about the varying label quality. What if we could somehow inform the model about the label quality? In this paper, we propose a semi-supervised learning method where we train two neural networks in a multi-task fashion: a "target network" and a "confidence network". The target network is optimized to perform a given task and is trained using a large set of unlabeled data that are weakly annotated. We propose to weight the gradient updates to the target network using the scores provided by the second confidence network, which is trained on a small amount of supervised data. Thus we avoid that the weight updates computed from noisy labels harm the quality of the target network model. We evaluate our learning strategy on two different tasks: document ranking and sentiment classification. The results demonstrate that our approach not only enhances the performance compared to the baselines but also speeds up the learning process from weak labels.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/30/2017

Learning to Learn from Weak Supervision by Full Supervision

In this paper, we propose a method for training neural networks when we ...
research
06/21/2018

Learning to Rank from Samples of Variable Quality

Training deep neural networks requires many training samples, but in pra...
research
11/08/2017

Fidelity-Weighted Learning

Training deep neural networks requires many training samples, but in pra...
research
04/12/2021

Self-Training with Weak Supervision

State-of-the-art deep neural networks require large-scale labeled traini...
research
02/06/2021

Jointly Improving Language Understanding and Generation with Quality-Weighted Weak Supervision of Automatic Labeling

Neural natural language generation (NLG) and understanding (NLU) models ...
research
06/29/2021

Semi-supervised learning with Bayesian Confidence Propagation Neural Network

Learning internal representations from data using no or few labels is us...
research
03/14/2017

Neural Graph Machines: Learning Neural Networks Using Graphs

Label propagation is a powerful and flexible semi-supervised learning te...

Please sign up or login with your details

Forgot password? Click here to reset