DeepAI AI Chat
Log In Sign Up

Distilling BERT for low complexity network training

05/13/2021
by   Bansidhar Mangalwedhekar, et al.
0

This paper studies the efficiency of transferring BERT learnings to low complexity models like BiLSTM, BiLSTM with attention and shallow CNNs using sentiment analysis on SST-2 dataset. It also compares the complexity of inference of the BERT model with these lower complexity models and underlines the importance of these techniques in enabling high performance NLP models on edge devices like mobiles, tablets and MCU development boards like Raspberry Pi etc. and enabling exciting new applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

01/10/2022

BERT for Sentiment Analysis: Pre-trained and Fine-Tuned Alternatives

BERT has revolutionized the NLP field by enabling transfer learning with...
11/20/2020

Fine-Tuning BERT for Sentiment Analysis of Vietnamese Reviews

Sentiment analysis is an important task in the field ofNature Language P...
09/06/2019

2-Local Hamiltonian with Low Complexity is QCMA

We prove that 2-Local Hamiltonian (2-LH) with Low Complexity problem is ...
09/19/2019

Coding for Optical Communications – Can We Approach the Shannon Limit With Low Complexity?

Approaching capacity with low complexity is a very challenging task. In ...
02/19/2023

Can ChatGPT Understand Too? A Comparative Study on ChatGPT and Fine-tuned BERT

Recently, ChatGPT has attracted great attention, as it can generate flue...
07/19/2018

Improving Simple Models with Confidence Profiles

In this paper, we propose a new method called ProfWeight for transferrin...