Testing pre-trained Transformer models for Lithuanian news clustering

04/03/2020
by   Lukas Stankevičius, et al.
0

A recent introduction of Transformer deep learning architecture made breakthroughs in various natural language processing tasks. However, non-English languages could not leverage such new opportunities with the English text pre-trained models. This changed with research focusing on multilingual models, where less-spoken languages are the main beneficiaries. We compare pre-trained multilingual BERT, XLM-R, and older learned text representation methods as encodings for the task of Lithuanian news clustering. Our results indicate that publicly available pre-trained multilingual Transformer models can be fine-tuned to surpass word vectors but still score much lower than specially trained doc2vec embeddings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/22/2020

mT5: A massively multilingual pre-trained text-to-text transformer

The recent "Text-to-Text Transfer Transformer" (T5) leveraged a unified ...
research
02/02/2023

idT5: Indonesian Version of Multilingual T5 Transformer

Indonesian language is spoken by almost 200 million people and is the 10...
research
02/25/2021

Are pre-trained text representations useful for multilingual and multi-dimensional language proficiency modeling?

Development of language proficiency models for non-native learners has b...
research
05/31/2022

Multilingual Transformers for Product Matching – Experiments and a New Benchmark in Polish

Product matching corresponds to the task of matching identical products ...
research
07/06/2023

Performance Comparison of Pre-trained Models for Speech-to-Text in Turkish: Whisper-Small and Wav2Vec2-XLS-R-300M

In this study, the performances of the Whisper-Small and Wav2Vec2-XLS-R-...
research
04/14/2023

ChatGPT: Applications, Opportunities, and Threats

Developed by OpenAI, ChatGPT (Conditional Generative Pre-trained Transfo...
research
11/02/2020

Introducing various Semantic Models for Amharic: Experimentation and Evaluation with multiple Tasks and Datasets

The availability of different pre-trained semantic models enabled the qu...

Please sign up or login with your details

Forgot password? Click here to reset