End to End Binarized Neural Networks for Text Classification

10/11/2020
by   Harshil Jain, et al.
0

Deep neural networks have demonstrated their superior performance in almost every Natural Language Processing task, however, their increasing complexity raises concerns. In particular, these networks require high expenses on computational hardware, and training budget is a concern for many. Even for a trained network, the inference phase can be too demanding for resource-constrained devices, thus limiting its applicability. The state-of-the-art transformer models are a vivid example. Simplifying the computations performed by a network is one way of relaxing the complexity requirements. In this paper, we propose an end to end binarized neural network architecture for the intent classification task. In order to fully utilize the potential of end to end binarization, both input representations (vector embeddings of tokens statistics) and the classifier are binarized. We demonstrate the efficiency of such architecture on the intent classification of short texts over three datasets and for text classification with a larger dataset. The proposed architecture achieves comparable to the state-of-the-art results on standard intent classification datasets while utilizing   20-40 lesser memory and training time. Furthermore, the individual components of the architecture, such as binarized vector embeddings of documents or binarized classifiers, can be used separately with not necessarily fully binary architectures.

READ FULL TEXT

page 2

page 8

research
06/24/2015

Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

Most tasks in natural language processing can be cast into question answ...
research
03/28/2019

Resilient Combination of Complementary CNN and RNN Features for Text Classification through Attention and Ensembling

State-of-the-art methods for text classification include several distinc...
research
04/03/2019

Rep the Set: Neural Networks for Learning Set Representations

In several domains, data objects can be decomposed into sets of simpler ...
research
05/16/2022

CONSENT: Context Sensitive Transformer for Bold Words Classification

We present CONSENT, a simple yet effective CONtext SENsitive Transformer...
research
10/31/2021

Text Classification for Task-based Source Code Related Questions

There is a key demand to automatically generate code for small tasks for...
research
07/18/2023

End-to-End Neural Network Training for Hyperbox-Based Classification

Hyperbox-based classification has been seen as a promising technique in ...
research
07/06/2018

JUMPER: Learning When to Make Classification Decisions in Reading

In early years, text classification is typically accomplished by feature...

Please sign up or login with your details

Forgot password? Click here to reset