Exploring Challenges of Deploying BERT-based NLP Models in Resource-Constrained Embedded Devices

04/23/2023
by   Souvika Sarkar, et al.
0

BERT-based neural architectures have established themselves as popular state-of-the-art baselines for many downstream NLP tasks. However, these architectures are data-hungry and consume a lot of memory and energy, often hindering their deployment in many real-time, resource-constrained applications. Existing lighter versions of BERT (eg. DistilBERT and TinyBERT) often cannot perform well on complex NLP tasks. More importantly, from a designer's perspective, it is unclear what is the "right" BERT-based architecture to use for a given NLP task that can strike the optimal trade-off between the resources available and the minimum accuracy desired by the end user. System engineers have to spend a lot of time conducting trial-and-error experiments to find a suitable answer to this question. This paper presents an exploratory study of BERT-based models under different resource constraints and accuracy budgets to derive empirical observations about this resource/accuracy trade-offs. Our findings can help designers to make informed choices among alternative BERT-based architectures for embedded systems, thus saving significant development time and effort.

READ FULL TEXT
research
10/25/2019

HUBERT Untangles BERT to Improve Transfer across NLP Tasks

We introduce HUBERT which combines the structured-representational power...
research
05/26/2020

Comparing BERT against traditional machine learning text classification

The BERT model has arisen as a popular state-of-the-art machine learning...
research
07/14/2021

Large-Scale News Classification using BERT Language Model: Spark NLP Approach

The rise of big data analytics on top of NLP increases the computational...
research
11/15/2022

When to Use What: An In-Depth Comparative Empirical Analysis of OpenIE Systems for Downstream Applications

Open Information Extraction (OpenIE) has been used in the pipelines of v...
research
08/31/2023

Dynamic nsNet2: Efficient Deep Noise Suppression with Early Exiting

Although deep learning has made strides in the field of deep noise suppr...
research
03/29/2020

User Generated Data: Achilles' heel of BERT

Pre-trained language models such as BERT are known to perform exceedingl...
research
11/03/2020

Towards Automated Anamnesis Summarization: BERT-based Models for Symptom Extraction

Professionals in modern healthcare systems are increasingly burdened by ...

Please sign up or login with your details

Forgot password? Click here to reset