PSG: Prompt-based Sequence Generation for Acronym Extraction

11/29/2021
by   Bin Li, et al.
0

Acronym extraction aims to find acronyms (i.e., short-forms) and their meanings (i.e., long-forms) from the documents, which is important for scientific document understanding (SDU@AAAI-22) tasks. Previous works are devoted to modeling this task as a paragraph-level sequence labeling problem. However, it lacks the effective use of the external knowledge, especially when the datasets are in a low-resource setting. Recently, the prompt-based method with the vast pre-trained language model can significantly enhance the performance of the low-resourced downstream tasks. In this paper, we propose a Prompt-based Sequence Generation (PSG) method for the acronym extraction task. Specifically, we design a template for prompting the extracted acronym texts with auto-regression. A position extraction algorithm is designed for extracting the position of the generated answers. The results on the acronym extraction of Vietnamese and Persian in a low-resource setting show that the proposed method outperforms all other competitive state-of-the-art (SOTA) methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/26/2022

AutoQGS: Auto-Prompt for Low-Resource Knowledge-based Question Generation from SPARQL

This study investigates the task of knowledge-based question generation ...
research
11/20/2019

Table-Of-Contents generation on contemporary documents

The generation of precise and detailed Table-Of-Contents (TOC) from a do...
research
05/10/2022

The Importance of Context in Very Low Resource Language Modeling

This paper investigates very low resource language model pretraining, wh...
research
12/20/2022

Pre-trained Language Models for Keyphrase Generation: A Thorough Empirical Study

Neural models that do not rely on pre-training have excelled in the keyp...
research
07/30/2023

A Knowledge-enhanced Two-stage Generative Framework for Medical Dialogue Information Extraction

This paper focuses on term-status pair extraction from medical dialogues...
research
12/30/2020

DEER: A Data Efficient Language Model for Event Temporal Reasoning

Pretrained language models (LMs) such as BERT, RoBERTa, and ELECTRA are ...
research
03/23/2022

Unified Structure Generation for Universal Information Extraction

Information extraction suffers from its varying targets, heterogeneous s...

Please sign up or login with your details

Forgot password? Click here to reset