Poolingformer: Long Document Modeling with Pooling Attention

05/10/2021
by   Hang Zhang, et al.
16

In this paper, we introduce a two-level attention schema, Poolingformer, for long document modeling. Its first level uses a smaller sliding window pattern to aggregate information from neighbors. Its second level employs a larger window to increase receptive fields with pooling attention to reduce both computational cost and memory consumption. We first evaluate Poolingformer on two long sequence QA tasks: the monolingual NQ and the multilingual TyDi QA. Experimental results show that Poolingformer sits atop three official leaderboards measured by F1, outperforming previous state-of-the-art models by 1.9 points (79.8 vs. 77.9) on NQ long answer, 1.9 points (79.5 vs. 77.6) on TyDi QA passage answer, and 1.6 points (67.6 vs. 66.0) on TyDi QA minimal answer. We further evaluate Poolingformer on a long sequence summarization task. Experimental results on the arXiv benchmark continue to demonstrate its superior performance.

READ FULL TEXT
research
05/03/2023

AttenWalker: Unsupervised Long-Document Question Answering via Attention-based Graph Walking

Annotating long-document question answering (long-document QA) pairs is ...
research
09/21/2022

Adapting Pretrained Text-to-Text Models for Long Text Sequences

We present an empirical study of adapting an existing pretrained text-to...
research
12/11/2019

Automatic Spanish Translation of the SQuAD Dataset for Multilingual Question Answering

Recently, multilingual question answering became a crucial research topi...
research
03/29/2018

Motion-Appearance Co-Memory Networks for Video Question Answering

Video Question Answering (QA) is an important task in understanding vide...
research
03/22/2021

Mitigating False-Negative Contexts in Multi-document QuestionAnswering with Retrieval Marginalization

Question Answering (QA) tasks requiring information from multiple docume...
research
10/09/2018

Answer Extraction in Question Answering using Structure Features and Dependency Principles

Question Answering (QA) research is a significant and challenging task i...
research
09/26/2021

QA-Align: Representing Cross-Text Content Overlap by Aligning Question-Answer Propositions

Multi-text applications, such as multi-document summarization, are typic...

Please sign up or login with your details

Forgot password? Click here to reset