DocPrompt: Large-scale continue pretrain for zero-shot and few-shot document question answering

08/21/2023
by   Sijin Wu, et al.
0

In this paper, we propose Docprompt for document question answering tasks with powerful zero-shot and few-shot performance. We proposed a novel weakly supervised data generation method, a novel multl-stage training method and a novel understanding model & generation model ensemble method. We achieved state-of-the-art performance on 4 document question answering tasks. This method greatly improves the delivery efficiency and model performance of document question answering customer projects, reducing annotation costs and labor costs. Our demo can be found at https://huggingface.co/spaces/PaddlePaddle/ERNIE-Layout.

READ FULL TEXT
research
06/01/2023

Layout and Task Aware Instruction Prompt for Zero-shot Document Image Question Answering

The pre-training-fine-tuning paradigm based on layout-aware multimodal p...
research
03/12/2021

Cooperative Learning of Zero-Shot Machine Reading Comprehension

Pretrained language models have significantly improved the performance o...
research
04/15/2022

Improving Passage Retrieval with Zero-Shot Question Generation

We propose a simple and effective re-ranking method for improving passag...
research
10/31/2022

Towards Zero-Shot and Few-Shot Table Question Answering using GPT-3

We present very early results on using GPT-3 to perform question answeri...
research
08/13/2021

Zero-shot Task Transfer for Invoice Extraction via Class-aware QA Ensemble

We present VESPA, an intentionally simple yet novel zero-shot system for...
research
04/21/2023

Information Extraction from Documents: Question Answering vs Token Classification in real-world setups

Research in Document Intelligence and especially in Document Key Informa...
research
05/24/2021

Few-Shot Upsampling for Protest Size Detection

We propose a new task and dataset for a common problem in social science...

Please sign up or login with your details

Forgot password? Click here to reset