Pseudo Labeling and Negative Feedback Learning for Large-scale Multi-label Domain Classification

03/08/2020
by   Joo-Kyung Kim, et al.
0

In large-scale domain classification, an utterance can be handled by multiple domains with overlapped capabilities. However, only a limited number of ground-truth domains are provided for each training utterance in practice while knowing as many as correct target labels is helpful for improving the model performance. In this paper, given one ground-truth domain for each training utterance, we regard domains consistently predicted with the highest confidences as additional pseudo labels for the training. In order to reduce prediction errors due to incorrect pseudo labels, we leverage utterances with negative system responses to decrease the confidences of the incorrectly predicted domains. Evaluating on user utterances from an intelligent conversational system, we show that the proposed approach significantly improves the performance of domain classification with hypothesis reranking.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/05/2021

Cycle Self-Training for Domain Adaptation

Mainstream approaches for unsupervised domain adaptation (UDA) learn dom...
research
09/04/2021

Error Detection in Large-Scale Natural Language Understanding Systems Using Transformer Models

Large-scale conversational assistants like Alexa, Siri, Cortana and Goog...
research
12/18/2018

Supervised Domain Enablement Attention for Personalized Domain Classification

In large-scale domain classification for natural language understanding,...
research
06/29/2018

Joint Learning of Domain Classification and Out-of-Domain Detection with Dynamic Class Weighting for Satisficing False Acceptance Rates

In domain classification for spoken dialog systems, correct detection of...
research
01/15/2021

In Defense of Pseudo-Labeling: An Uncertainty-Aware Pseudo-label Selection Framework for Semi-Supervised Learning

The recent research in semi-supervised learning (SSL) is mostly dominate...
research
11/24/2019

Enhancing Out-Of-Domain Utterance Detection with Data Augmentation Based on Word Embeddings

For most intelligent assistant systems, it is essential to have a mechan...
research
07/02/2020

Climbing the WOL: Training for Cheaper Inference

Efficient inference for wide output layers (WOLs) is an essential yet ch...

Please sign up or login with your details

Forgot password? Click here to reset