Transformer-GCRF: Recovering Chinese Dropped Pronouns with General Conditional Random Fields

10/07/2020
by   Jingxuan Yang, et al.
0

Pronouns are often dropped in Chinese conversations and recovering the dropped pronouns is important for NLP applications such as Machine Translation. Existing approaches usually formulate this as a sequence labeling task of predicting whether there is a dropped pronoun before each token and its type. Each utterance is considered to be a sequence and labeled independently. Although these approaches have shown promise, labeling each utterance independently ignores the dependencies between pronouns in neighboring utterances. Modeling these dependencies is critical to improving the performance of dropped pronoun recovery. In this paper, we present a novel framework that combines the strength of Transformer network with General Conditional Random Fields (GCRF) to model the dependencies between pronouns in neighboring utterances. Results on three Chinese conversation datasets show that the Transformer-GCRF model outperforms the state-of-the-art dropped pronoun recovery models. Exploratory analysis also demonstrates that the GCRF did help to capture the dependencies between pronouns in neighboring utterances, thus contributes to the performance improvements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2019

Who did They Respond to? Conversation Structure Modeling using Masked Hierarchical Transformer

Conversation structure is useful for both understanding the nature of co...
research
11/10/2020

Neural Latent Dependency Model for Sequence Labeling

Sequence labeling is a fundamental problem in machine learning, natural ...
research
06/07/2021

A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech

In this paper, we present a neural model for joint dropped pronoun recov...
research
05/07/2016

Neural Recovery Machine for Chinese Dropped Pronoun

Dropped pronouns (DPs) are ubiquitous in pro-drop languages like Chinese...
research
02/21/2023

Generic Dependency Modeling for Multi-Party Conversation

To model the dependencies between utterances in multi-party conversation...
research
10/27/2022

Conversation Disentanglement with Bi-Level Contrastive Learning

Conversation disentanglement aims to group utterances into detached sess...
research
02/21/2020

Guider l'attention dans les modeles de sequence a sequence pour la prediction des actes de dialogue

The task of predicting dialog acts (DA) based on conversational dialog i...

Please sign up or login with your details

Forgot password? Click here to reset