Re-Evaluating GermEval17 Using German Pre-Trained Language Models

02/24/2021
by   M. Aßenmacher, et al.
0

The lack of a commonly used benchmark data set (collection) such as (Super-)GLUE (Wang et al., 2018, 2019) for the evaluation of non-English pre-trained language models is a severe shortcoming of current English-centric NLP-research. It concentrates a large part of the research on English, neglecting the uncertainty when transferring conclusions found for the English language to other languages. We evaluate the performance of the German and multilingual BERT-based models currently available via the huggingface transformers library on the four tasks of the GermEval17 workshop. We compare them to pre-BERT architectures (Wojatzki et al., 2017; Schmitt et al., 2018; Attia et al., 2018) as well as to an ELMo-based architecture (Biesialska et al., 2020) and a BERT-based approach (Guhr et al., 2020). The observed improvements are put in relation to those for similar tasks and similar models (pre-BERT vs. BERT-based) for the English language in order to draw tentative conclusions about whether the observed improvements are transferable to German or potentially other related languages.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2020

GottBERT: a pure German Language Model

Lately, pre-trained language models advanced the field of natural langua...
research
09/20/2021

Model Bias in NLP – Application to Hate Speech Classification

This document sums up our results forthe NLP lecture at ETH in the sprin...
research
07/14/2023

How Different Is Stereotypical Bias Across Languages?

Recent studies have demonstrated how to assess the stereotypical bias in...
research
02/27/2020

A Primer in BERTology: What we know about how BERT works

Transformer-based models are now widely used in NLP, but we still do not...
research
12/04/2020

Pre-trained language models as knowledge bases for Automotive Complaint Analysis

Recently it has been shown that large pre-trained language models like B...
research
05/16/2019

Latent Universal Task-Specific BERT

This paper describes a language representation model which combines the ...
research
10/04/2021

DeepA2: A Modular Framework for Deep Argument Analysis with Pretrained Neural Text2Text Language Models

In this paper, we present and implement a multi-dimensional, modular fra...

Please sign up or login with your details

Forgot password? Click here to reset