Towards Deconfounding the Influence of Subject's Demographic Characteristics in Question Answering

04/15/2021
by   Maharshi Gor, et al.
0

Question Answering (QA) tasks are used as benchmarks of general machine intelligence. Therefore, robust QA evaluation is critical, and metrics should indicate how models will answer any question. However, major QA datasets have skewed distributions over gender, profession, and nationality. Despite that skew, models generalize – we find little evidence that accuracy is lower for people based on gender or nationality. Instead, there is more variation in question topic and question ambiguity. Adequately accessing the generalization of QA systems requires more representative datasets.

READ FULL TEXT
research
06/30/2022

Modern Question Answering Datasets and Benchmarks: A Survey

Question Answering (QA) is one of the most important natural language pr...
research
06/07/2017

Question Answering and Question Generation as Dual Tasks

We study the problem of joint question answering (QA) and question gener...
research
09/12/2019

Finding Generalizable Evidence by Learning to Convince Q A Models

We propose a system that finds the strongest supporting evidence for a g...
research
09/21/2021

What Would it Take to get Biomedical QA Systems into Practice?

Medical question answering (QA) systems have the potential to answer cli...
research
04/09/2022

Extending the Scope of Out-of-Domain: Examining QA models in multiple subdomains

Past works that investigate out-of-domain performance of QA systems have...
research
01/17/2017

Community Question Answering Platforms vs. Twitter for Predicting Characteristics of Urban Neighbourhoods

In this paper, we investigate whether text from a Community Question Ans...
research
04/21/2022

Benchmarking Answer Verification Methods for Question Answering-Based Summarization Evaluation Metrics

Question answering-based summarization evaluation metrics must automatic...

Please sign up or login with your details

Forgot password? Click here to reset