Scaled-up Discovery of Latent Concepts in Deep NLP Models

08/20/2023
by   Majd Hawasly, et al.
0

Pre-trained language models (pLMs) learn intricate patterns and contextual dependencies via unsupervised learning on vast text data, driving breakthroughs across NLP tasks. Despite these achievements, these models remain black boxes, necessitating research into understanding their decision-making processes. Recent studies explore representation analysis by clustering latent spaces within pre-trained models. However, these approaches are limited in terms of scalability and the scope of interpretation because of high computation costs of clustering algorithms. This study focuses on comparing clustering algorithms for the purpose of scaling encoded concept discovery of representations from pLMs. Specifically, we compare three algorithms in their capacity to unveil the encoded concepts through their alignment to human-defined ontologies: Agglomerative Hierarchical Clustering, Leaders Algorithm, and K-Means Clustering. Our results show that K-Means has the potential to scale to very large datasets, allowing rich latent concept discovery, both on the word and phrase level.

READ FULL TEXT
research
05/22/2023

Can LLMs facilitate interpretation of pre-trained language models?

Work done to uncover the knowledge encoded within pre-trained language m...
research
06/27/2022

Analyzing Encoded Concepts in Transformer Language Models

We propose a novel framework ConceptX, to analyze how latent concepts ar...
research
11/12/2022

ConceptX: A Framework for Latent Concept Analysis

The opacity of deep neural networks remains a challenge in deploying sol...
research
10/23/2022

On the Transformation of Latent Space in Fine-Tuned NLP Models

We study the evolution of latent space in fine-tuned NLP models. Differe...
research
04/29/2020

General Purpose Text Embeddings from Pre-trained Language Models for Scalable Inference

The state of the art on many NLP tasks is currently achieved by large pr...
research
02/03/2023

Towards Few-Shot Identification of Morality Frames using In-Context Learning

Data scarcity is a common problem in NLP, especially when the annotation...
research
08/31/2021

It's not Rocket Science : Interpreting Figurative Language in Narratives

Figurative language is ubiquitous in English. Yet, the vast majority of ...

Please sign up or login with your details

Forgot password? Click here to reset