Biomedical Data-to-Text Generation via Fine-Tuning Transformers

09/03/2021
by   Ruslan Yermakov, et al.
0

Data-to-text (D2T) generation in the biomedical domain is a promising - yet mostly unexplored - field of research. Here, we apply neural models for D2T generation to a real-world dataset consisting of package leaflets of European medicines. We show that fine-tuned transformers are able to generate realistic, multisentence text from data in the biomedical domain, yet have important limitations. We also release a new dataset (BioLeaflets) for benchmarking D2T generation models in the biomedical domain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/26/2021

Fine-tuning Pretrained Language Models with Label Attention for Explainable Biomedical Text Classification

The massive growth of digital biomedical data is making biomedical text ...
research
02/04/2022

Transformers and the representation of biomedical background knowledge

BioBERT and BioMegatron are Transformers models adapted for the biomedic...
research
02/06/2018

Texygen: A Benchmarking Platform for Text Generation Models

We introduce Texygen, a benchmarking platform to support research on ope...
research
11/22/2022

A Large-Scale Dataset for Biomedical Keyphrase Generation

Keyphrase generation is the task consisting in generating a set of words...
research
06/07/2023

Evaluation of ChatGPT on Biomedical Tasks: A Zero-Shot Comparison with Fine-Tuned Generative Transformers

ChatGPT is a large language model developed by OpenAI. Despite its impre...
research
07/01/2023

More for Less: Compact Convolutional Transformers Enable Robust Medical Image Classification with Limited Data

Transformers are very powerful tools for a variety of tasks across domai...
research
12/24/2021

Spoiler in a Textstack: How Much Can Transformers Help?

This paper presents our research regarding spoiler detection in reviews....

Please sign up or login with your details

Forgot password? Click here to reset