HiJoNLP at SemEval-2022 Task 2: Detecting Idiomaticity of Multiword Expressions using Multilingual Pretrained Language Models

05/27/2022
by   Minghuan Tan, et al.
0

This paper describes an approach to detect idiomaticity only from the contextualized representation of a MWE over multilingual pretrained language models. Our experiments find that larger models are usually more effective in idiomaticity detection. However, using a higher layer of the model may not guarantee a better performance. In multilingual scenarios, the convergence of different languages are not consistent and rich-resource languages have big advantages over other languages.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2022

Factual Consistency of Multilingual Pretrained Language Models

Pretrained language models can be queried for factual knowledge, with po...
research
04/11/2022

Adapting BigScience Multilingual Model to Unseen Languages

We benchmark different strategies of adding new languages (German and Ko...
research
10/13/2020

X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained Language Models

Language models (LMs) have proven surprisingly successful at capturing f...
research
09/13/2021

Wine is Not v i n. – On the Compatibility of Tokenizations Across Languages

The size of the vocabulary is a central design choice in large pretraine...
research
04/22/2023

Romanian Multiword Expression Detection Using Multilingual Adversarial Training and Lateral Inhibition

Multiword expressions are a key ingredient for developing large-scale an...
research
04/18/2023

UniMax: Fairer and more Effective Language Sampling for Large-Scale Multilingual Pretraining

Pretrained multilingual large language models have typically used heuris...
research
08/19/2020

BabelEnconding at SemEval-2020 Task 3: Contextual Similarity as a Combination of Multilingualism and Language Models

This paper describes the system submitted by our team (BabelEnconding) t...

Please sign up or login with your details

Forgot password? Click here to reset