TEAGS: Time-aware Text Embedding Approach to Generate Subgraphs

07/06/2019
by   Saeid Hosseini, et al.
0

Contagions (e.g. virus, gossip) spread over the nodes in propagation graphs. We can use the temporal and textual data of the nodes to compute the edge weights and then generate subgraphs with highly relevant nodes. This is beneficial to many applications. Yet, challenges abound. First, the propagation pattern between each pair of nodes may change by time. Second, not always the same contagion propagates. Hence, the state-of-the-art text mining approaches including topic-modeling cannot effectively compute the edge weights. Third, since the propagation is affected by time, the word-word co-occurrence patterns may differ in various temporal dimensions, that can decrease the effectiveness of word embedding approaches. We argue that multi-aspect temporal dimensions (hour, day, etc) should be considered to better calculate the correlation weights between the nodes. In this work, we devise a novel framework that on the one hand, integrates a neural network based time-aware word embedding component to construct the word vectors through multiple temporal facets, and on the other hand, uses a temporal generative model to compute the weights. Subsequently, we propose a Max-Heap Graph cutting algorithm to generate subgraphs. We validate our model through comprehensive experiments on real-world datasets. The results show that our model can retrieve the subgraphs more effective than other rivals and the temporal dynamics should be noticed both in word embedding and propagation processes.

READ FULL TEXT

page 5

page 14

research
07/06/2019

TEALS: Time-aware Text Embedding Approach to Leverage Subgraphs

Given a graph over which the contagions (e.g. virus, gossip) propagate, ...
research
01/11/2023

Word-Graph2vec: An efficient word embedding approach on word co-occurrence graph using random walk sampling

Word embedding has become ubiquitous and is widely used in various text ...
research
05/09/2018

LearningWord Embeddings for Low-resource Languages by PU Learning

Word embedding is a key component in many downstream applications in pro...
research
10/27/2019

SoulMate: Short-text author linking through Multi-aspect temporal-textual embedding

Linking authors of short-text contents has important usages in many appl...
research
07/10/2023

Source-Aware Embedding Training on Heterogeneous Information Networks

Heterogeneous information networks (HINs) have been extensively applied ...
research
07/17/2022

Representation Learning of Image Schema

Image schema is a recurrent pattern of reasoning where one entity is map...
research
07/08/2020

Mining Dense Subgraphs with Similar Edges

When searching for interesting structures in graphs, it is often importa...

Please sign up or login with your details

Forgot password? Click here to reset