Smaller Text Classifiers with Discriminative Cluster Embeddings

06/23/2019
by   Mingda Chen, et al.
0

Word embedding parameters often dominate overall model sizes in neural methods for natural language processing. We reduce deployed model sizes of text classifiers by learning a hard word clustering in an end-to-end manner. We use the Gumbel-Softmax distribution to maximize over the latent clustering while minimizing the task loss. We propose variations that selectively assign additional parameters to words, which further improves accuracy while still remaining parameter-efficient.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/07/2023

An Investigation on Word Embedding Offset Clustering as Relationship Classification

Vector representations obtained from word embedding are the source of ma...
research
11/12/2019

word2ket: Space-efficient Word Embeddings inspired by Quantum Entanglement

Deep learning natural language processing models often use vector word e...
research
12/28/2019

Learning Numeral Embeddings

Word embedding is an essential building block for deep learning methods ...
research
03/05/2021

Overcoming Poor Word Embeddings with Word Definitions

Modern natural language understanding models depend on pretrained subwor...
research
03/22/2019

An end-to-end Neural Network Framework for Text Clustering

The unsupervised text clustering is one of the major tasks in natural la...
research
01/19/2018

Investigating the Working of Text Classifiers

Text classification is one of the most widely studied task in natural la...
research
10/03/2021

Subtractive mountain clustering algorithm applied to a chatbot to assist elderly people in medication intake

Errors in medication intake among elderly people are very common. One of...

Please sign up or login with your details

Forgot password? Click here to reset