Federated pretraining and fine tuning of BERT using clinical notes from multiple silos

02/20/2020
by   Dianbo Liu, et al.
0

Large scale contextual representation models, such as BERT, have significantly advanced natural language processing (NLP) in recently years. However, in certain area like healthcare, accessing diverse large scale text data from multiple institutions is extremely challenging due to privacy and regulatory reasons. In this article, we show that it is possible to both pretrain and fine tune BERT models in a federated manner using clinical texts from different silos without moving the data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/14/2019

Two-stage Federated Phenotyping and Patient Representation Learning

A large percentage of medical information is in unstructured text format...
research
11/12/2020

An Interpretable End-to-end Fine-tuning Approach for Long Clinical Text

Unstructured clinical text in EHRs contains crucial information for appl...
research
12/09/2021

Transferring BERT-like Transformers' Knowledge for Authorship Verification

The task of identifying the author of a text spans several decades and w...
research
06/28/2022

The NLP Sandbox: an efficient model-to-data system to enable federated and unbiased evaluation of clinical NLP models

Objective The evaluation of natural language processing (NLP) models for...
research
05/26/2020

BERT-XML: Large Scale Automated ICD Coding Using BERT Pretraining

Clinical interactions are initially recorded and documented in free text...
research
10/12/2020

Layer-wise Guided Training for BERT: Learning Incrementally Refined Document Representations

Although BERT is widely used by the NLP community, little is known about...
research
08/15/2019

M-BERT: Injecting Multimodal Information in the BERT Structure

Multimodal language analysis is an emerging research area in natural lan...

Please sign up or login with your details

Forgot password? Click here to reset