Pre-trained Language Model Based Active Learning for Sentence Matching

10/12/2020
by   Guirong Bai, et al.
0

Active learning is able to significantly reduce the annotation cost for data-driven techniques. However, previous active learning approaches for natural language processing mainly depend on the entropy-based uncertainty criterion, and ignore the characteristics of natural language. In this paper, we propose a pre-trained language model based active learning approach for sentence matching. Differing from previous active learning, it can provide linguistic criteria to measure instances and help select more efficient instances for annotation. Experiments demonstrate our approach can achieve greater accuracy with fewer labeled training instances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/20/2021

Active Learning for Sequence Tagging with Deep Pre-trained Models and Bayesian Uncertainty Estimates

Annotating training data for sequence tagging tasks is usually very time...
research
07/31/2023

A Pre-trained Data Deduplication Model based on Active Learning

In the era of big data, the issue of data quality has become increasingl...
research
04/17/2020

Active Sentence Learning by Adversarial Uncertainty Sampling in Discrete Space

In this paper, we focus on reducing the labeled data size for sentence l...
research
06/22/2011

Acquiring Word-Meaning Mappings for Natural Language Interfaces

This paper focuses on a system, WOLFIE (WOrd Learning From Interpreted E...
research
09/26/2017

Active Learning amidst Logical Knowledge

Structured prediction is ubiquitous in applications of machine learning ...
research
12/02/2015

Active Learning for Delineation of Curvilinear Structures

Many recent delineation techniques owe much of their increased effective...
research
04/08/2021

Deep Indexed Active Learning for Matching Heterogeneous Entity Representations

Given two large lists of records, the task in entity resolution (ER) is ...

Please sign up or login with your details

Forgot password? Click here to reset