Gendered Ambiguous Pronouns Shared Task: Boosting Model Confidence by Evidence Pooling

06/03/2019
by   Sandeep Attree, et al.
0

This paper presents a strong set of results for resolving gendered ambiguous pronouns on the Gendered Ambiguous Pronouns shared task. The model presented here draws upon the strengths of state-of-the-art language and coreference resolution models, and introduces a novel evidence-based deep learning architecture. Injecting evidence from the coreference models compliments the base architecture, and analysis shows that the model is not hindered by their weaknesses, specifically gender bias. The modularity and simplicity of the architecture make it very easy to extend for further improvement and applicable to other NLP problems. Evaluation on GAP test data results in a state-of-the-art performance at 92.5 to the human performance of 96.6 placed 1st in the Kaggle competition, winning by a significant lead. The code is available at https://github.com/sattree/gap.

READ FULL TEXT
research
06/09/2019

Gendered Pronoun Resolution using BERT and an extractive question answering formulation

The resolution of ambiguous pronouns is a longstanding challenge in Natu...
research
05/18/2023

Improving Toponym Resolution with Better Candidate Generation, Transformer-based Reranking, and Two-Stage Resolution

Geocoding is the task of converting location mentions in text into struc...
research
09/24/2020

Type B Reflexivization as an Unambiguous Testbed for Multilingual Multi-Task Gender Bias

The one-sided focus on English in previous studies of gender bias in NLP...
research
10/11/2018

Mind the GAP: A Balanced Corpus of Gendered Ambiguous Pronouns

Coreference resolution is an important task for natural language underst...
research
05/24/2023

Are Pre-trained Language Models Useful for Model Ensemble in Chinese Grammatical Error Correction?

Model ensemble has been in widespread use for Grammatical Error Correcti...
research
10/21/2022

SimANS: Simple Ambiguous Negatives Sampling for Dense Text Retrieval

Sampling proper negatives from a large document pool is vital to effecti...
research
07/07/2021

Can Transformer Models Measure Coherence In Text? Re-Thinking the Shuffle Test

The Shuffle Test is the most common task to evaluate whether NLP models ...

Please sign up or login with your details

Forgot password? Click here to reset