GraphFormers: GNN-nested Language Models for Linked Text Representation

05/06/2021
by   Junhan Yang, et al.
13

Linked text representation is critical for many intelligent web applications, such as online advertisement and recommender systems. Recent breakthroughs on pretrained language models and graph neural networks facilitate the development of corresponding techniques. However, the existing works mainly rely on cascaded model structures: the texts are independently encoded by language models at first, and the textual embeddings are further aggregated by graph neural networks. We argue that the neighbourhood information is insufficiently utilized within the above process, which restricts the representation quality. In this work, we propose GraphFormers, where graph neural networks are nested alongside each transformer layer of the language models. On top of the above architecture, the linked texts will iteratively extract neighbourhood information for the enhancement of their own semantics. Such an iterative workflow gives rise to more effective utilization of neighbourhood information, which contributes to the representation quality. We further introduce an adaptation called unidirectional GraphFormers, which is much more efficient and comparably effective; and we leverage a pretraining strategy called the neighbourhood-aware masked language modeling to enhance the training effect. We perform extensive experiment studies with three large-scale linked text datasets, whose results verify the effectiveness of our proposed methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2022

Efficient and effective training of language and graph neural network models

Can we combine heterogenous graph structure with text to learn high-qual...
research
05/20/2022

Heterformer: A Transformer Architecture for Node Representation Learning on Heterogeneous Text-Rich Networks

We study node representation learning on heterogeneous text-rich network...
research
09/27/2021

Graph Neural Networks for Recommender Systems: Challenges, Methods, and Directions

Recommender system is one of the most important information services on ...
research
10/26/2022

Learning on Large-scale Text-attributed Graphs via Variational Inference

This paper studies learning on text-attributed graphs (TAGs), where each...
research
03/06/2022

Graph Neural Network Enhanced Language Models for Efficient Multilingual Text Classification

Online social media works as a source of various valuable and actionable...
research
01/25/2022

Whose Language Counts as High Quality? Measuring Language Ideologies in Text Data Selection

Language models increasingly rely on massive web dumps for diverse text ...

Please sign up or login with your details

Forgot password? Click here to reset