Efficient Learning for Undirected Topic Models

06/24/2015
by   Jiatao Gu, et al.
0

Replicated Softmax model, a well-known undirected topic model, is powerful in extracting semantic representations of documents. Traditional learning strategies such as Contrastive Divergence are very inefficient. This paper provides a novel estimator to speed up the learning based on Noise Contrastive Estimate, extended for documents of variant lengths and weighted inputs. Experiments on two benchmarks show that the new estimator achieves great learning efficiency and high accuracy on document retrieval and classification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2023

Efficient Document Embeddings via Self-Contrastive Bregman Divergence Learning

Learning quality document embeddings is a fundamental problem in natural...
research
03/04/2020

Contrastive estimation reveals topic posterior information to linear models

Contrastive learning is an approach to representation learning that util...
research
07/04/2012

Mining Associated Text and Images with Dual-Wing Harmoniums

We propose a multi-wing harmonium model for mining multimedia data that ...
research
04/19/2023

Shuffle Divide: Contrastive Learning for Long Text

We propose a self-supervised learning method for long text documents bas...
research
09/26/2013

Modeling Documents with Deep Boltzmann Machines

We introduce a Deep Boltzmann Machine model suitable for modeling and ex...
research
08/17/2018

Learning Supervised Topic Models for Classification and Regression from Crowds

The growing need to analyze large collections of documents has led to gr...

Please sign up or login with your details

Forgot password? Click here to reset