Crisis Domain Adaptation Using Sequence-to-sequence Transformers

10/15/2021
by   Congcong Wang, et al.
0

User-generated content (UGC) on social media can act as a key source of information for emergency responders in crisis situations. However, due to the volume concerned, computational techniques are needed to effectively filter and prioritise this content as it arises during emerging events. In the literature, these techniques are trained using annotated content from previous crises. In this paper, we investigate how this prior knowledge can be best leveraged for new crises by examining the extent to which crisis events of a similar type are more suitable for adaptation to new events (cross-domain adaptation). Given the recent successes of transformers in various language processing tasks, we propose CAST: an approach for Crisis domain Adaptation leveraging Sequence-to-sequence Transformers. We evaluate CAST using two major crisis-related message classification datasets. Our experiments show that our CAST-based best run without using any target data achieves the state of the art performance in both in-domain and cross-domain contexts. Moreover, CAST is particularly effective in one-to-one cross-domain adaptation when trained with a larger language model. In many-to-one adaptation where multiple crises are jointly used as the source domain, CAST further improves its performance. In addition, we find that more similar events are more likely to bring better adaptation performance whereas fine-tuning using dissimilar events does not help for adaptation. To aid reproducibility, we open source our code to the community.

READ FULL TEXT

page 1

page 8

research
05/26/2023

Coping with low data availability for social media crisis message categorisation

During crisis situations, social media allows people to quickly share in...
research
08/09/2016

Multi-task Domain Adaptation for Sequence Tagging

Many domain adaptation approaches rely on learning cross domain shared r...
research
10/22/2020

Knowledge Distillation for BERT Unsupervised Domain Adaptation

A pre-trained language model, BERT, has brought significant performance ...
research
11/17/2014

Joint cross-domain classification and subspace learning for unsupervised adaptation

Domain adaptation aims at adapting the knowledge acquired on a source do...
research
03/24/2021

DRANet: Disentangling Representation and Adaptation Networks for Unsupervised Cross-Domain Adaptation

In this paper, we present DRANet, a network architecture that disentangl...
research
08/27/2023

Domain-Specificity Inducing Transformers for Source-Free Domain Adaptation

Conventional Domain Adaptation (DA) methods aim to learn domain-invarian...
research
08/06/2021

Distilling Transformers for Neural Cross-Domain Search

Pre-trained transformers have recently clinched top spots in the gamut o...

Please sign up or login with your details

Forgot password? Click here to reset