Unsupervised Pre-training for Biomedical Question Answering

09/27/2020
by   Vaishnavi Kommaraju, et al.
0

We explore the suitability of unsupervised representation learning methods on biomedical text – BioBERT, SciBERT, and BioSentVec – for biomedical question answering. To further improve unsupervised representations for biomedical QA, we introduce a new pre-training task from unlabeled data designed to reason about biomedical entities in the context. Our pre-training method consists of corrupting a given context by randomly replacing some mention of a biomedical entity with a random entity mention and then querying the model with the correct entity mention in order to locate the corrupted part of the context. This de-noising task enables the model to learn good representations from abundant, unlabeled biomedical text that helps QA tasks and minimizes the train-test mismatch between the pre-training task and the downstream QA tasks by requiring the model to predict spans. Our experiments show that pre-training BioBERT on the proposed pre-training task significantly boosts performance and outperforms the previous best model from the 7th BioASQ Task 7b-Phase B challenge.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/26/2022

Contextual embedding and model weighting by fusing domain knowledge on Biomedical Question Answering

Biomedical Question Answering aims to obtain an answer to the given ques...
research
07/10/2023

Enhancing Biomedical Text Summarization and Question-Answering: On the Utility of Domain-Specific Pre-Training

Biomedical summarization requires large datasets to train for text gener...
research
04/12/2023

CLIP-Guided Vision-Language Pre-training for Question Answering in 3D Scenes

Training models to apply linguistic knowledge and visual concepts from 2...
research
09/21/2021

Relation-Guided Pre-Training for Open-Domain Question Answering

Answering complex open-domain questions requires understanding the laten...
research
10/11/2022

Mixed-modality Representation Learning and Pre-training for Joint Table-and-Text Retrieval in OpenQA

Retrieving evidences from tabular and textual resources is essential for...
research
05/30/2019

Unsupervised pre-training helps to conserve views from input distribution

We investigate the effects of the unsupervised pre-training method under...
research
06/15/2021

Question Answering Infused Pre-training of General-Purpose Contextualized Representations

This paper proposes a pre-training objective based on question answering...

Please sign up or login with your details

Forgot password? Click here to reset