As research on machine translation moves to translating text beyond the
...
Automatic evaluation of machine translation (MT) is a critical tool driv...
Kendall's tau is frequently used to meta-evaluate how well machine
trans...
The acquisition of high-quality human annotations through crowdsourcing
...
There is significant interest in developing evaluation metrics which
acc...
We introduce Repro, an open-source library which aims at improving the
r...
How reliably an automatic summarization evaluation metric replicates hum...
Question answering-based summarization evaluation metrics must automatic...
In this work, we propose a method for incorporating question-answering (...
The quality of a summarization evaluation metric is quantified by calcul...
Reference-based metrics such as ROUGE or BERTScore evaluate the content
...
Recently, there has been growing interest in using question-answering (Q...
We present SacreROUGE, an open-source library for using and developing
s...