On the interaction of automatic evaluation and task framing in headline style transfer

01/05/2021
by   Lorenzo De Mattei, et al.
10

An ongoing debate in the NLG community concerns the best way to evaluate systems, with human evaluation often being considered the most reliable method, compared to corpus-based metrics. However, tasks involving subtle textual differences, such as style transfer, tend to be hard for humans to perform. In this paper, we propose an evaluation method for this task based on purposely-trained classifiers, showing that it better reflects system differences than traditional metrics such as BLEU and ROUGE.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2021

A Review of Human Evaluation for Style Transfer

This paper reviews and summarizes human evaluation practices described i...
research
05/24/2023

Don't Take This Out of Context! On the Need for Contextual Models and Evaluations for Stylistic Rewriting

Most existing stylistic text rewriting methods operate on a sentence lev...
research
03/17/2018

Dear Sir or Madam, May I introduce the YAFC Corpus: Corpus, Benchmarks and Metrics for Formality Style Transfer

Style transfer is the task of automatically transforming a piece of text...
research
08/19/2019

Style Transfer for Texts: to Err is Human, but Error Margins Matter

This paper shows that standard assessment methodology for style transfer...
research
03/17/2018

Dear Sir or Madam, May I introduce the GYAFC Dataset: Corpus, Benchmarks and Metrics for Formality Style Transfer

Style transfer is the task of automatically transforming a piece of text...
research
08/19/2019

Style Transfer for Texts: Retrain, Report Errors, Compare with Rewrites

This paper shows that standard assessment methodology for style transfer...
research
10/08/2019

Prose for a Painting

Painting captions are often dry and simplistic which motivates us to des...

Please sign up or login with your details

Forgot password? Click here to reset