SML: a new Semantic Embedding Alignment Transformer for efficient cross-lingual Natural Language Inference

03/17/2021
by   Javier Huertas-Tato, et al.
26

The ability of Transformers to perform with precision a variety of tasks such as question answering, Natural Language Inference (NLI) or summarising, have enable them to be ranked as one of the best paradigms to address this kind of tasks at present. NLI is one of the best scenarios to test these architectures, due to the knowledge required to understand complex sentences and established a relation between a hypothesis and a premise. Nevertheless, these models suffer from incapacity to generalise to other domains or difficulties to face multilingual scenarios. The leading pathway in the literature to address these issues involve designing and training extremely large architectures, which leads to unpredictable behaviours and to establish barriers which impede broad access and fine tuning. In this paper, we propose a new architecture, siamese multilingual transformer (SML), to efficiently align multilingual embeddings for Natural Language Inference. SML leverages siamese pre-trained multi-lingual transformers with frozen weights where the two input sentences attend each other to later be combined through a matrix alignment method. The experimental results carried out in this paper evidence that SML allows to reduce drastically the number of trainable parameters while still achieving state-of-the-art performance.

READ FULL TEXT

page 5

page 6

research
05/22/2023

Enhancing Cross-lingual Natural Language Inference by Soft Prompting with Multilingual Verbalizer

Cross-lingual natural language inference is a fundamental problem in cro...
research
05/15/2023

Measuring Cross-Lingual Transferability of Multilingual Transformers on Sentence Classification

Recent studies have exhibited remarkable capabilities of pre-trained mul...
research
04/18/2017

Baselines and test data for cross-lingual inference

Research in natural language inference is currently exclusive to English...
research
05/12/2022

Lifting the Curse of Multilinguality by Pre-training Modular Transformers

Multilingual pre-trained models are known to suffer from the curse of mu...
research
04/18/2022

Exploring Dimensionality Reduction Techniques in Multilingual Transformers

Both in scientific literature and in industry,, Semantic and context-awa...
research
02/17/2022

Designing Effective Sparse Expert Models

Scale has opened new frontiers in natural language processing – but at a...
research
05/07/2021

∂-Explainer: Abductive Natural Language Inference via Differentiable Convex Optimization

Constrained optimization solvers with Integer Linear programming (ILP) h...

Please sign up or login with your details

Forgot password? Click here to reset