Hierarchical Neural Language Models for Joint Representation of Streaming Documents and their Content

06/28/2016
by   Nemanja Djuric, et al.
0

We consider the problem of learning distributed representations for documents in data streams. The documents are represented as low-dimensional vectors and are jointly learned with distributed vector representations of word tokens using a hierarchical framework with two embedded neural language models. In particular, we exploit the context of documents in streams and use one of the language models to model the document sequences, and the other to model word sequences within them. The models learn continuous vector representations for both word tokens and documents such that semantically similar documents and words are close in a common vector space. We discuss extensions to our model, which can be applied to personalized recommendation and social relationship mining by adding further user layers to the hierarchy, thus learning user-specific vectors to represent individual preferences. We validated the learned representations on a public movie rating data set from MovieLens, as well as on a large-scale Yahoo News data comprising three months of user activity logs collected on Yahoo servers. The results indicate that the proposed model can learn useful representations of both documents and word tokens, outperforming the current state-of-the-art by a large margin.

READ FULL TEXT
research
08/09/2017

Neural Vector Spaces for Unsupervised Information Retrieval

We propose the Neural Vector Space Model (NVSM), a method that learns re...
research
12/30/2022

On the Inconsistencies of Conditionals Learned by Masked Language Models

Learning to predict masked tokens in a sequence has been shown to be a p...
research
05/11/2020

Multidirectional Associative Optimization of Function-Specific Word Representations

We present a neural framework for learning associations between interrel...
research
03/07/2020

Discovering linguistic (ir)regularities in word embeddings through max-margin separating hyperplanes

We experiment with new methods for learning how related words are positi...
research
04/27/2015

Document Classification by Inversion of Distributed Language Representations

There have been many recent advances in the structure and measurement of...
research
03/09/2023

On the Robustness of Text Vectorizers

A fundamental issue in natural language processing is the robustness of ...
research
06/05/2023

Leveraging Large Language Models for Topic Classification in the Domain of Public Affairs

The analysis of public affairs documents is crucial for citizens as it p...

Please sign up or login with your details

Forgot password? Click here to reset