On the Limitations of Sociodemographic Adaptation with Transformers

08/01/2022
by   Chia-Chien Hung, et al.
0

Sociodemographic factors (e.g., gender or age) shape our language. Previous work showed that incorporating specific sociodemographic factors can consistently improve performance for various NLP tasks in traditional NLP models. We investigate whether these previous findings still hold with state-of-the-art pretrained Transformers. We use three common specialization methods proven effective for incorporating external knowledge into pretrained Transformers (e.g., domain-specific or geographic knowledge). We adapt the language representations for the sociodemographic dimensions of gender and age, using continuous language modeling and dynamic multi-task learning for adaptation, where we couple language modeling with the prediction of a sociodemographic class. Our results when employing a multilingual model show substantial performance gains across four languages (English, German, French, and Danish). These findings are in line with the results of previous work and hold promise for successful sociodemographic specialization. However, controlling for confounding factors like domain and language shows that, while sociodemographic adaptation does improve downstream performance, the gains do not always solely stem from sociodemographic knowledge. Our results indicate that sociodemographic specialization, while very important, is still an unresolved problem in NLP.

READ FULL TEXT

page 8

page 9

research
10/13/2022

Can Demographic Factors Improve Text Classification? Revisiting Demographic Adaptation in the Age of Transformers

Demographic factors (e.g., gender or age) shape our language. Previous w...
research
12/31/2020

Verb Knowledge Injection for Multilingual Event Processing

In parallel to their overwhelming success across NLP tasks, language abi...
research
03/16/2022

Geographic Adaptation of Pretrained Language Models

Geographic linguistic features are commonly used to improve the performa...
research
03/14/2022

Efficient Language Modeling with Sparse all-MLP

All-MLP architectures have attracted increasing interest as an alternati...
research
08/27/2023

Domain-Specificity Inducing Transformers for Source-Free Domain Adaptation

Conventional Domain Adaptation (DA) methods aim to learn domain-invarian...
research
10/11/2021

Multi-Task Learning for Situated Multi-Domain End-to-End Dialogue Systems

Task-oriented dialogue systems have been a promising area in the NLP fie...
research
07/31/2022

Neural Knowledge Bank for Pretrained Transformers

The ability of pretrained Transformers to remember factual knowledge is ...

Please sign up or login with your details

Forgot password? Click here to reset