Can ChatGPT Detect Intent? Evaluating Large Language Models for Spoken Language Understanding

05/22/2023
by   Mutian He, et al.
0

Recently, large pretrained language models have demonstrated strong language understanding capabilities. This is particularly reflected in their zero-shot and in-context learning abilities on downstream tasks through prompting. To assess their impact on spoken language understanding (SLU), we evaluate several such models like ChatGPT and OPT of different sizes on multiple benchmarks. We verify the emergent ability unique to the largest models as they can reach intent classification accuracy close to that of supervised models with zero or few shots on various languages given oracle transcripts. By contrast, the results for smaller models fitting a single GPU fall far behind. We note that the error cases often arise from the annotation scheme of the dataset; responses from ChatGPT are still reasonable. We show, however, that the model is worse at slot filling, and its performance is sensitive to ASR errors, suggesting serious challenges for the application of those textual models on SLU.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/26/2023

Zero-Shot Slot and Intent Detection in Low-Resource Languages

Intent detection and slot filling are critical tasks in spoken and natur...
research
02/01/2020

Fine-Tuning BERT for Schema-Guided Zero-Shot Dialogue State Tracking

We present our work on Track 4 in the Dialogue System Technology Challen...
research
01/05/2023

HIT-SCIR at MMNLU-22: Consistency Regularization for Multilingual Spoken Language Understanding

Multilingual spoken language understanding (SLU) consists of two sub-tas...
research
12/03/2019

Fast Intent Classification for Spoken Language Understanding

Spoken Language Understanding (SLU) systems consist of several machine l...
research
08/23/2023

Devising and Detecting Phishing: Large Language Models vs. Smaller Human Models

AI programs, built using large language models, make it possible to auto...
research
08/30/2023

ToddlerBERTa: Exploiting BabyBERTa for Grammar Learning and Language Understanding

We present ToddlerBERTa, a BabyBERTa-like language model, exploring its ...
research
06/20/2019

One-vs-All Models for Asynchronous Training: An Empirical Analysis

Any given classification problem can be modeled using multi-class or One...

Please sign up or login with your details

Forgot password? Click here to reset