Accelerating Natural Language Understanding in Task-Oriented Dialog

06/05/2020
by   Ojas Ahuja, et al.
0

Task-oriented dialog models typically leverage complex neural architectures and large-scale, pre-trained Transformers to achieve state-of-the-art performance on popular natural language understanding benchmarks. However, these models frequently have in excess of tens of millions of parameters, making them impossible to deploy on-device where resource-efficiency is a major concern. In this work, we show that a simple convolutional model compressed with structured pruning achieves largely comparable results to BERT on ATIS and Snips, with under 100K parameters. Moreover, we perform acceleration experiments on CPUs, where we observe our multi-task model predicts intents and slots nearly 63x faster than even DistilBERT.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2019

Reweighted Proximal Pruning for Large-Scale Language Representation

Recently, pre-trained language representation flourishes as the mainstay...
research
12/30/2020

Robustness Testing of Language Understanding in Dialog Systems

Most language understanding models in dialog systems are trained on a sm...
research
10/28/2019

A Simple but Effective BERT Model for Dialog State Tracking on Resource-Limited Systems

In a task-oriented dialog system, the goal of dialog state tracking (DST...
research
06/12/2020

A Generative Model for Joint Natural Language Understanding and Generation

Natural language understanding (NLU) and natural language generation (NL...
research
05/21/2022

Calibration of Natural Language Understanding Models with Venn–ABERS Predictors

Transformers, currently the state-of-the-art in natural language underst...
research
11/30/2020

Extreme Model Compression for On-device Natural Language Understanding

In this paper, we propose and experiment with techniques for extreme com...
research
07/19/2018

Statistical Model Compression for Small-Footprint Natural Language Understanding

In this paper we investigate statistical model compression applied to na...

Please sign up or login with your details

Forgot password? Click here to reset