A Novel Multi-Task Learning Approach for Context-Sensitive Compound Type Identification in Sanskrit

08/22/2022
by   Jivnesh Sandhan, et al.
0

The phenomenon of compounding is ubiquitous in Sanskrit. It serves for achieving brevity in expressing thoughts, while simultaneously enriching the lexical and structural formation of the language. In this work, we focus on the Sanskrit Compound Type Identification (SaCTI) task, where we consider the problem of identifying semantic relations between the components of a compound word. Earlier approaches solely rely on the lexical information obtained from the components and ignore the most crucial contextual and syntactic information useful for SaCTI. However, the SaCTI task is challenging primarily due to the implicitly encoded context-sensitive semantic relation between the compound components. Thus, we propose a novel multi-task learning architecture which incorporates the contextual information and enriches the complementary syntactic information using morphological tagging and dependency parsing as two auxiliary tasks. Experiments on the benchmark datasets for SaCTI show 6.1 points (Accuracy) and 7.7 points (F1-score) absolute gain compared to the state-of-the-art system. Further, our multi-lingual experiments demonstrate the efficacy of the proposed architecture in English and Marathi languages.The code and datasets are publicly available at https://github.com/ashishgupta2598/SaCTI

READ FULL TEXT
research
01/05/2021

PhoNLP: A joint multi-task learning model for Vietnamese part-of-speech tagging, named entity recognition and dependency parsing

We present the first multi-task learning model – named PhoNLP – for join...
research
10/12/2020

Contextual Modulation for Relation-Level Metaphor Identification

Identifying metaphors in text is very challenging and requires comprehen...
research
08/09/2021

Benchmarking Multi-Task Learning for Sentiment Analysis and Offensive Language Identification in Under-Resourced Dravidian Languages

To obtain extensive annotated data for under-resourced languages is chal...
research
02/12/2021

A Little Pretraining Goes a Long Way: A Case Study on Dependency Parsing Task for Low-resource Morphologically Rich Languages

Neural dependency parsing has achieved remarkable performance for many d...
research
01/27/2022

Systematic Investigation of Strategies Tailored for Low-Resource Settings for Sanskrit Dependency Parsing

Existing state of the art approaches for Sanskrit Dependency Parsing (SD...
research
06/03/2016

Exploiting Multi-typed Treebanks for Parsing with Deep Multi-task Learning

Various treebanks have been released for dependency parsing. Despite tha...
research
03/09/2023

A Syntax-Guided Multi-Task Learning Approach for Turducken-Style Code Generation

Due to the development of pre-trained language models, automated code ge...

Please sign up or login with your details

Forgot password? Click here to reset