Building a Hebrew Semantic Role Labeling Lexical Resource from Parallel Movie Subtitles

05/17/2020
by   Ben Eyal, et al.
0

We present a semantic role labeling resource for Hebrew built semi-automatically through annotation projection from English. This corpus is derived from the multilingual OpenSubtitles dataset and includes short informal sentences, for which reliable linguistic annotations have been computed. We provide a fully annotated version of the data including morphological analysis, dependency syntax and semantic role labeling in both FrameNet and PropBank styles. Sentences are aligned between English and Hebrew, both sides include full annotations and the explicit mapping from the English arguments to the Hebrew ones. We train a neural SRL model on this Hebrew resource exploiting the pre-trained multilingual BERT transformer model, and provide the first available baseline model for Hebrew SRL as a reference point. The code we provide is generic and can be adapted to other languages to bootstrap SRL resources.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2020

X-SRL: A Parallel Cross-Lingual Semantic Role Labeling Dataset

Even though SRL is researched for many languages, major improvements hav...
research
05/29/2018

Polyglot Semantic Role Labeling

Previous approaches to multilingual semantic dependency parsing treat la...
research
08/29/2019

Translate and Label! An Encoder-Decoder Approach for Cross-lingual Semantic Role Labeling

We propose a Cross-lingual Encoder-Decoder model that simultaneously tra...
research
04/07/2017

Adposition Supersenses v2

This document describes an inventory of 50 semantic labels designed to c...
research
01/04/2021

Transformers and Transfer Learning for Improving Portuguese Semantic Role Labeling

Semantic Role Labeling (SRL) is a core Natural Language Processing task....
research
08/05/2019

Semantic Role Labeling with Associated Memory Network

Semantic role labeling (SRL) is a task to recognize all the predicate-ar...
research
08/31/2019

Small and Practical BERT Models for Sequence Labeling

We propose a practical scheme to train a single multilingual sequence la...

Please sign up or login with your details

Forgot password? Click here to reset