The Utility of General Domain Transfer Learning for Medical Language Tasks

02/16/2020
by   Daniel Ranti, et al.
13

The purpose of this study is to analyze the efficacy of transfer learning techniques and transformer-based models as applied to medical natural language processing (NLP) tasks, specifically radiological text classification. We used 1,977 labeled head CT reports, from a corpus of 96,303 total reports, to evaluate the efficacy of pretraining using general domain corpora and a combined general and medical domain corpus with a bidirectional representations from transformers (BERT) model for the purpose of radiological text classification. Model performance was benchmarked to a logistic regression using bag-of-words vectorization and a long short-term memory (LSTM) multi-label multi-class classification model, and compared to the published literature in medical text classification. The BERT models using either set of pretrained checkpoints outperformed the logistic regression model, achieving sample-weighted average F1-scores of 0.87 and 0.87 for the general domain model and the combined general and biomedical-domain model. General text transfer learning may be a viable technique to generate state-of-the-art results within medical NLP tasks on radiological corpora, outperforming other deep models such as LSTMs. The efficacy of pretraining and transformer-based models could serve to facilitate the creation of groundbreaking NLP models in the uniquely challenging data environment of medical text.

READ FULL TEXT
research
09/02/2021

LegaLMFiT: Efficient Short Legal Text Classification with LSTM Language Model Pre-Training

Large Transformer-based language models such as BERT have led to broad p...
research
03/13/2023

Transformer-based approaches to Sentiment Detection

The use of transfer learning methods is largely responsible for the pres...
research
10/24/2020

Large Scale Legal Text Classification Using Transformer Models

Large multi-label text classification is a challenging Natural Language ...
research
07/02/2020

Bidirectional Encoder Representations from Transformers (BERT): A sentiment analysis odyssey

The purpose of the study is to investigate the relative effectiveness of...
research
06/28/2023

Multi-Site Clinical Federated Learning using Recursive and Attentive Models and NVFlare

The prodigious growth of digital health data has precipitated a mounting...
research
03/21/2023

Fine-tuning ClimateBert transformer with ClimaText for the disclosure analysis of climate-related financial risks

In recent years there has been a growing demand from financial agents, e...
research
10/31/2022

Improving Cause-of-Death Classification from Verbal Autopsy Reports

In many lower-and-middle income countries including South Africa, data a...

Please sign up or login with your details

Forgot password? Click here to reset