Semi-Automated Labeling of Requirement Datasets for Relation Extraction

09/05/2021
by   Jeremias Bohn, et al.
0

Creating datasets manually by human annotators is a laborious task that can lead to biased and inhomogeneous labels. We propose a flexible, semi-automatic framework for labeling data for relation extraction. Furthermore, we provide a dataset of preprocessed sentences from the requirements engineering domain, including a set of automatically created as well as hand-crafted labels. In our case study, we compare the human and automatic labels and show that there is a substantial overlap between both annotations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/06/2023

FinRED: A Dataset for Relation Extraction in Financial Domain

Relation extraction models trained on a source domain cannot be applied ...
research
04/14/2022

FREDA: Flexible Relation Extraction Data Annotation

To effectively train accurate Relation Extraction models, sufficient and...
research
10/17/2020

Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction

Distant supervision has been a widely used method for neural relation ex...
research
04/30/2020

Revisiting Unsupervised Relation Extraction

Unsupervised relation extraction (URE) extracts relations between named ...
research
04/28/2020

A Practical Framework for Relation Extraction with Noisy Labels Based on Doubly Transitional Loss

Either human annotation or rule based automatic labeling is an effective...
research
10/17/2019

Exploring Semi-Automatic Map Labeling

Label placement in maps is a very challenging task that is critical for ...
research
04/28/2022

What do You Mean by Relation Extraction? A Survey on Datasets and Study on Scientific Relation Classification

Over the last five years, research on Relation Extraction (RE) witnessed...

Please sign up or login with your details

Forgot password? Click here to reset