Techniques to Improve Q A Accuracy with Transformer-based models on Large Complex Documents

09/26/2020
by   Chejui Liao, et al.
0

This paper discusses the effectiveness of various text processing techniques, their combinations, and encodings to achieve a reduction of complexity and size in a given text corpus. The simplified text corpus is sent to BERT (or similar transformer based models) for question and answering and can produce more relevant responses to user queries. This paper takes a scientific approach to determine the benefits and effectiveness of various techniques and concludes a best-fit combination that produces a statistically significant improvement in accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/16/2020

Delaying Interaction Layers in Transformer-based Encoders for Efficient Open Domain Question Answering

Open Domain Question Answering (ODQA) on a large-scale corpus of documen...
research
08/11/2023

Identification of the Relevance of Comments in Codes Using Bag of Words and Transformer Based Models

The Forum for Information Retrieval (FIRE) started a shared task this ye...
research
05/25/2020

Dialect Text Normalization to Normative Standard Finnish

We compare different LSTMs and transformer models in terms of their effe...
research
08/22/2020

FAT ALBERT: Finding Answers in Large Texts using Semantic Similarity Attention Layer based on BERT

Machine based text comprehension has always been a significant research ...
research
02/14/2020

Stress Test Evaluation of Transformer-based Models in Natural Language Understanding Tasks

There has been significant progress in recent years in the field of Natu...
research
05/31/2021

Corpus-Based Paraphrase Detection Experiments and Review

Paraphrase detection is important for a number of applications, includin...
research
11/04/2022

BERT for Long Documents: A Case Study of Automated ICD Coding

Transformer models have achieved great success across many NLP problems....

Please sign up or login with your details

Forgot password? Click here to reset