Long-Tailed Question Answering in an Open World

05/11/2023
by   Yi Dai, et al.
0

Real-world data often have an open long-tailed distribution, and building a unified QA model supporting various tasks is vital for practical QA applications. However, it is non-trivial to extend previous QA approaches since they either require access to seen tasks of adequate samples or do not explicitly model samples from unseen tasks. In this paper, we define Open Long-Tailed QA (OLTQA) as learning from long-tailed distributed data and optimizing performance over seen and unseen QA tasks. We propose an OLTQA model that encourages knowledge sharing between head, tail and unseen tasks, and explicitly mines knowledge from a large pre-trained language model (LM). Specifically, we organize our model through a pool of fine-grained components and dynamically combine these components for an input to facilitate knowledge sharing. A retrieve-then-rerank frame is further introduced to select in-context examples, which guild the LM to generate text that express knowledge for QA tasks. Moreover, a two-stage training approach is introduced to pre-train the framework by knowledge distillation (KD) from the LM and then jointly train the frame and a QA model through an adaptive mutual KD method. On a large-scale OLTQA dataset we curate from 43 existing QA datasets, our model consistently outperforms the state-of-the-art. We release the code and data at <https://github.com/AlibabaResearch/DAMO-ConvAI/tree/main/oltqa>.

READ FULL TEXT
research
08/31/2022

Lifelong Learning for Question Answering with Hierarchical Prompts

QA models with lifelong learning (LL) abilities are important for practi...
research
05/11/2023

Domain Incremental Lifelong Learning in an Open World

Lifelong learning (LL) is an important ability for NLP models to learn n...
research
05/02/2023

Huatuo-26M, a Large-scale Chinese Medical QA Dataset

In this paper, we release a largest ever medical Question Answering (QA)...
research
09/21/2021

Relation-Guided Pre-Training for Open-Domain Question Answering

Answering complex open-domain questions requires understanding the laten...
research
08/17/2022

Open Long-Tailed Recognition in a Dynamic World

Real world data often exhibits a long-tailed and open-ended (with unseen...
research
05/25/2022

LEPUS: Prompt-based Unsupervised Multi-hop Reranking for Open-domain QA

We study unsupervised multi-hop reranking for multi-hop QA (MQA) with op...
research
02/23/2023

Extracting Victim Counts from Text

Decision-makers in the humanitarian sector rely on timely and exact info...

Please sign up or login with your details

Forgot password? Click here to reset