Improving Neural Topic Models using Knowledge Distillation

10/05/2020
by   Alexander Hoyle, et al.
0

Topic models are often used to identify human-interpretable topics to help make sense of large document collections. We use knowledge distillation to combine the best attributes of probabilistic topic models and pretrained transformers. Our modular method can be straightforwardly applied with any neural topic model to improve topic quality, which we demonstrate using two models having disparate architectures, obtaining state-of-the-art topic coherence. We show that our adaptable framework not only improves performance in the aggregate over all estimated topics, as is commonly reported, but also in head-to-head comparisons of aligned topics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/27/2023

Improving Neural Topic Models with Wasserstein Knowledge Distillation

Topic modeling is a dominant method for exploring document collections o...
research
04/13/2023

G2T: A Simple but Effective Framework for Topic Modeling based on Pretrained Language Model and Community Detection

It has been reported that clustering-based topic models, which cluster h...
research
06/19/2020

Neural Topic Modeling with Continual Lifelong Learning

Lifelong learning has recently attracted attention in building machine l...
research
03/27/2023

Improving Contextualized Topic Models with Negative Sampling

Topic modeling has emerged as a dominant method for exploring large docu...
research
04/17/2021

Multi-source Neural Topic Modeling in Multi-view Embedding Spaces

Though word embeddings and topics are complementary representations, sev...
research
07/31/2017

Combining Thesaurus Knowledge and Probabilistic Topic Models

In this paper we present the approach of introducing thesaurus knowledge...
research
08/01/2016

Labeling Topics with Images using Neural Networks

Topics generated by topic models are usually represented by lists of t t...

Please sign up or login with your details

Forgot password? Click here to reset