CogAlign: Learning to Align Textual Neural Representations to Cognitive Language Processing Signals

06/10/2021
by   Yuqi Ren, et al.
0

Most previous studies integrate cognitive language processing signals (e.g., eye-tracking or EEG data) into neural models of natural language processing (NLP) just by directly concatenating word embeddings with cognitive features, ignoring the gap between the two modalities (i.e., textual vs. cognitive) and noise in cognitive features. In this paper, we propose a CogAlign approach to these issues, which learns to align textual neural representations to cognitive features. In CogAlign, we use a shared encoder equipped with a modality discriminator to alternatively encode textual and cognitive inputs to capture their differences and commonalities. Additionally, a text-aware attention mechanism is proposed to detect task-related information and to avoid using noise in cognitive features. Experimental results on three NLP tasks, namely named entity recognition, sentiment analysis and relation extraction, show that CogAlign achieves significant improvements with multiple cognitive features over state-of-the-art models on public datasets. Moreover, our model is able to transfer cognitive information to other datasets that do not have any cognitive processing signals.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2019

Advancing NLP with Cognitive Language Processing Signals

When we read, our brain processes language and generates cognitive proce...
research
12/16/2021

Bridging between Cognitive Processing Signals and Linguistic Features via a Unified Attentional Network

Cognitive processing signals can be used to improve natural language pro...
research
09/19/2019

CogniVal: A Framework for Cognitive Word Embedding Evaluation

An interesting method of evaluating word representations is by how much ...
research
06/09/2020

Human brain activity for machine attention

Cognitively inspired NLP leverages human-derived data to teach machines ...
research
02/03/2021

Memorization vs. Generalization: Quantifying Data Leakage in NLP Performance Evaluation

Public datasets are often used to evaluate the efficacy and generalizabi...
research
02/11/2023

Synthesizing Human Gaze Feedback for Improved NLP Performance

Integrating human feedback in models can improve the performance of natu...
research
06/08/2016

On the Place of Text Data in Lifelogs, and Text Analysis via Semantic Facets

Current research in lifelog data has not paid enough attention to analys...

Please sign up or login with your details

Forgot password? Click here to reset