Logical reasoning of text is an important ability that requires understa...
Answering open-domain questions requires world knowledge about in-contex...
A common thread of retrieval-augmented methods in the existing literatur...
Large language models (LLMs) show impressive abilities via few-shot
prom...
Knowledge-intensive tasks, such as open-domain question answering (QA),
...
The goal of this work is to build flexible video-language models that ca...
Vision-language (V+L) pretraining models have achieved great success in
...
We initiate the first empirical study on the use of MLP architectures fo...
Most of today's AI systems focus on using self-attention mechanisms and
...
Large-scale pre-trained language models have achieved tremendous success...
Vision-and-language (VL) pre-training has proven to be highly effective ...
Commonsense reasoning (CSR) requires the model to be equipped with gener...
Pre-trained language models (PLMs) aim to learn universal language
repre...
Current Open-Domain Question Answering (ODQA) model paradigm often conta...
Large-scale transformer-based pre-training has recently revolutionized
v...
Vision-and-language pre-training has achieved impressive success in lear...
Lottery Ticket Hypothesis raises keen attention to identifying sparse
tr...
Multimodal pre-training has propelled great advancement in
vision-and-la...
Deep, heavily overparameterized language models such as BERT, XLNet and ...
Deep neural network based question answering (QA) models are neither rob...
In this paper, we propose Cross-Thought, a novel approach to pre-trainin...
Pre-trained neural abstractive summarization systems have dominated
extr...
Large-scale language models such as BERT have achieved state-of-the-art
...
Existing language model compression methods mostly use a simple L2 loss ...
Transformer has become ubiquitous in the deep learning field. One of the...
Existing approaches to real-time question answering (RTQA) rely on learn...
Large-scale cross-lingual language models (LM), such as mBERT, Unicoder ...
Transformer has been successfully applied to many natural language proce...
In this paper, we present Hierarchical Graph Network (HGN) for multi-hop...
Multiple-Choice Reading Comprehension (MCRC) requires the model to read ...
Many state-of-the-art neural models for NLP are heavily parameterized an...
This paper tackles the problem of reading comprehension over long narrat...
Commonsense reasoning is fundamental to natural language understanding. ...
Multi-choice reading comprehension is a challenging task, which involves...
A popular recent approach to answering open-domain questions is to first...
In recent years researchers have achieved considerable success applying
...
Many NLP tasks including machine comprehension, answer selection and tex...
Machine comprehension of text is an important problem in natural languag...
Natural language inference (NLI) is a fundamentally important task in na...