Summarizing Indian Languages using Multilingual Transformers based Models

03/29/2023
by   Dhaval Taunk, et al.
0

With the advent of multilingual models like mBART, mT5, IndicBART etc., summarization in low resource Indian languages is getting a lot of attention now a days. But still the number of datasets is low in number. In this work, we (Team HakunaMatata) study how these multilingual models perform on the datasets which have Indian languages as source and target text while performing summarization. We experimented with IndicBART and mT5 models to perform the experiments and report the ROUGE-1, ROUGE-2, ROUGE-3 and ROUGE-4 scores as a performance metric.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/25/2021

XL-Sum: Large-Scale Multilingual Abstractive Summarization for 44 Languages

Contemporary works on abstractive text summarization have focused primar...
research
04/26/2018

Lessons from the Bible on Modern Topics: Low-Resource Multilingual Topic Model Evaluation

Multilingual topic models enable document analysis across languages thro...
research
07/15/2023

Multilingual Adapter-based Knowledge Aggregation on Code Summarization for Low-Resource Languages

Multilingual fine-tuning (of a multilingual Pre-trained Language Model) ...
research
10/17/2021

Predicting the Performance of Multilingual NLP Models

Recent advancements in NLP have given us models like mBERT and XLMR that...
research
05/24/2022

MaskEval: Weighted MLM-Based Evaluation for Text Summarization and Simplification

In text summarization and simplification, system outputs must be evaluat...
research
06/07/2023

Echoes from Alexandria: A Large Resource for Multilingual Book Summarization

In recent years, research in text summarization has mainly focused on th...
research
10/11/2022

Shapley Head Pruning: Identifying and Removing Interference in Multilingual Transformers

Multilingual transformer-based models demonstrate remarkable zero and fe...

Please sign up or login with your details

Forgot password? Click here to reset