ALL-IN-1: Short Text Classification with One Model for All Languages

10/26/2017
by   Barbara Plank, et al.
0

We present ALL-IN-1, a simple model for multilingual text classification that does not require any parallel data. It is based on a traditional Support Vector Machine classifier exploiting multilingual word embeddings and character n-grams. Our model is simple, easily extendable yet very effective, overall ranking 1st (out of 12 teams) in the IJCNLP 2017 shared task on customer feedback analysis in four languages: English, French, Japanese and Spanish.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2019

Expanding the Text Classification Toolbox with Cross-Lingual Embeddings

Most work in text classification and Natural Language Processing (NLP) f...
research
02/18/2023

RetVec: Resilient and Efficient Text Vectorizer

This paper describes RetVec, a resilient multilingual embedding scheme d...
research
03/28/2023

Model and Evaluation: Towards Fairness in Multilingual Text Classification

Recently, more and more research has focused on addressing bias in text ...
research
11/29/2019

A Multi-cascaded Deep Model for Bilingual SMS Classification

Most studies on text classification are focused on the English language....
research
06/05/2018

Understanding Meanings in Multilingual Customer Feedback

Understanding and being able to react to customer feedback is the most f...
research
08/19/2023

Optimizing Multi-Class Text Classification: A Diverse Stacking Ensemble Framework Utilizing Transformers

Customer reviews play a crucial role in assessing customer satisfaction,...
research
05/15/2023

Taxi1500: A Multilingual Dataset for Text Classification in 1500 Languages

While natural language processing tools have been developed extensively ...

Please sign up or login with your details

Forgot password? Click here to reset