Rethinking Self-Supervision Objectives for Generalizable Coherence Modeling

10/14/2021
by   Prathyusha Jwalapuram, et al.
5

Although large-scale pre-trained neural models have shown impressive performances in a variety of tasks, their ability to generate coherent text that appropriately models discourse phenomena is harder to evaluate and less understood. Given the claims of improved text generation quality across various systems, we consider the coherence evaluation of machine generated text to be one of the principal applications of coherence models that needs to be investigated. We explore training data and self-supervision objectives that result in a model that generalizes well across tasks and can be used off-the-shelf to perform such evaluations. Prior work in neural coherence modeling has primarily focused on devising new architectures, and trained the model to distinguish coherent and incoherent text through pairwise self-supervision on the permuted documents task. We instead use a basic model architecture and show significant improvements over state of the art within the same training regime. We then design a harder self-supervision objective by increasing the ratio of negative samples within a contrastive learning setup, and enhance the model further through automatic hard negative mining coupled with a large global negative queue encoded by a momentum encoder. We show empirically that increasing the density of negative samples improves the basic model, and using a global negative queue further improves and stabilizes the model while training with hard negative samples. We evaluate the coherence model on task-independent test sets that resemble real-world use cases and show significant improvements in coherence evaluations of downstream applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2016

Neural Net Models for Open-Domain Discourse Coherence

Discourse coherence is strongly associated with text quality, making it ...
research
11/01/2018

A bird's-eye view on coherence, and a worm's-eye view on cohesion

Generating coherent and cohesive long-form texts is a challenging proble...
research
04/30/2020

CohEval: Benchmarking Coherence Models

Although coherence modeling has come a long way in developing novel mode...
research
09/22/2022

Learning to Write with Coherence From Negative Examples

Coherence is one of the critical factors that determine the quality of w...
research
05/14/2018

Discourse Coherence in the Wild: A Dataset, Evaluation and Methods

To date there has been very little work on assessing discourse coherence...
research
03/18/2022

DEAM: Dialogue Coherence Evaluation using AMR-based Semantic Manipulations

Automatic evaluation metrics are essential for the rapid development of ...
research
12/20/2022

CoCo: Coherence-Enhanced Machine-Generated Text Detection Under Data Limitation With Contrastive Learning

Machine-Generated Text (MGT) detection, a task that discriminates MGT fr...

Please sign up or login with your details

Forgot password? Click here to reset