Improving Social Meaning Detection with Pragmatic Masking and Surrogate Fine-Tuning

08/01/2021
by   Chiyu Zhang, et al.
0

Masked language models (MLMs) are pretrained with a denoising objective that, while useful, is in a mismatch with the objective of downstream fine-tuning. We propose pragmatic masking and surrogate fine-tuning as two strategies that exploit social cues to drive pre-trained representations toward a broad set of concepts useful for a wide class of social meaning tasks. To test our methods, we introduce a new benchmark of 15 different Twitter datasets for social meaning detection. Our methods achieve 2.34 while outperforming other transfer learning methods such as multi-task learning and domain-specific language models pretrained on large datasets. With only 5 of training data (severely few-shot), our methods enable an impressive 68.74 average F1, and we observe promising results in a zero-shot setting involving six datasets from three different languages.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro