This paper proposes a new method, OFA-OCR, to transfer multimodal pretra...
Generalist models, which are capable of performing diverse multi-modal t...
Vision-and-language (V-L) tasks require the system to understand both vi...
Recently, attention based models have been used extensively in many
sequ...
Investigating better ways to reuse the released pre-trained language mod...
Contrastive Language-Image Pre-training (CLIP) has demonstrated great
po...
Pre-trained models have achieved excellent performance on the dialogue t...
As many fine-tuned pre-trained language models (PLMs) with promising
per...
The conventional wisdom behind learning deep classification models is to...
The class imbalance problem, as an important issue in learning node
repr...
Previous studies demonstrate DNNs' vulnerability to adversarial examples...
Video captioning combines video understanding and language generation.
D...
Skip connection, is a widely-used technique to improve the performance a...
Recent studies have revealed a security threat to natural language proce...
Neural dialogue models suffer from low-quality responses when interacted...
Dynamic early exiting aims to accelerate pre-trained language models' (P...
Pre-trained self-supervised models such as BERT have achieved striking
s...
Human dialogues are scenario-based and appropriate responses generally r...
Collaborative learning has successfully applied knowledge transfer to gu...
We argue that the vulnerability of model parameters is of crucial value ...
In sequence-to-sequence learning, the attention mechanism has been a gre...
Recently, attention-based encoder-decoder models have been used extensiv...
Self-attention based Transformer has demonstrated the state-of-the-art
p...
Training deep neural networks requires intricate initialization and care...
Considering event structure information has proven helpful in text-based...
Chinese word segmentation (CWS) is a fundamental step of Chinese natural...
Unsupervised text style transfer aims to alter text styles while preserv...
Neural network learning is typically slow since backpropagation needs to...
In image-grounded text generation, fine-grained representations of the i...
Automatic evaluation of semantic rationality is an important yet challen...
The encode-decoder framework has shown recent success in image captionin...
This paper explores a new natural language processing task, review-drive...
Most of the Neural Machine Translation (NMT) models are based on the
seq...
Narrative story generation is a challenging problem because it demands t...
Huge numbers of new words emerge every day, leading to a great need for
...
A great proportion of sequence-to-sequence (Seq2Seq) models for Neural
M...
The goal of sentiment-to-sentiment "translation" is to change the underl...
Abstractive text summarization is a highly difficult problem, and the
se...
Text summarization and sentiment classification both aim to capture the ...
Relation classification is an important semantic processing task in the ...
Most recent approaches use the sequence-to-sequence model for paraphrase...
Most recent approaches use the sequence-to-sequence model for paraphrase...
Existing text generation methods tend to produce repeated and "boring"
e...
Web 2.0 has brought with it numerous user-produced data revealing one's
...
In the training of transition-based dependency parsers, an oracle is use...
Recent systems on structured prediction focus on increasing the level of...
We propose a simple yet effective technique to simplify the training and...
We propose a method, called Label Embedding Network, which can learn lab...
As traditional neural network consumes a significant amount of computing...
We propose a simple yet effective technique for neural network learning....