Automatic punctuation restoration with BERT models

01/18/2021
by   Attila Nagy, et al.
0

We present an approach for automatic punctuation restoration with BERT models for English and Hungarian. For English, we conduct our experiments on Ted Talks, a commonly used benchmark for punctuation restoration, while for Hungarian we evaluate our models on the Szeged Treebank dataset. Our best models achieve a macro-averaged F_1-score of 79.8 in English and 82.2 in Hungarian. Our code is publicly available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2021

Diacritics Restoration using BERT with Analysis on Czech language

We propose a new architecture for diacritics restoration based on contex...
research
10/23/2021

PhoMT: A High-Quality and Large-Scale Benchmark Dataset for Vietnamese-English Machine Translation

We introduce a high-quality and large-scale Vietnamese-English parallel ...
research
08/24/2023

A Small and Fast BERT for Chinese Medical Punctuation Restoration

In clinical dictation, utterances after automatic speech recognition (AS...
research
12/10/2022

Punctuation Restoration for Singaporean Spoken Languages: English, Malay, and Mandarin

This paper presents the work of restoring punctuation for ASR transcript...
research
04/06/2021

SERRANT: a syntactic classifier for English Grammatical Error Types

SERRANT is a system and code for automatic classification of English gra...
research
07/10/2019

Dunhuang Grottoes Painting Dataset and Benchmark

This document introduces the background and the usage of the Dunhuang Gr...
research
07/10/2019

Dunhuang Grotto Painting Dataset and Benchmark

This document introduces the background and the usage of the Dunhuang Gr...

Please sign up or login with your details

Forgot password? Click here to reset