On the Relationship Between RNN Hidden State Vectors and Semantic Ground Truth

06/29/2023
by   Edi Muskardin, et al.
0

We examine the assumption that the hidden-state vectors of recurrent neural networks (RNNs) tend to form clusters of semantically similar vectors, which we dub the clustering hypothesis. While this hypothesis has been assumed in the analysis of RNNs in recent years, its validity has not been studied thoroughly on modern neural network architectures. We examine the clustering hypothesis in the context of RNNs that were trained to recognize regular languages. This enables us to draw on perfect ground-truth automata in our evaluation, against which we can compare the RNN's accuracy and the distribution of the hidden-state vectors. We start with examining the (piecewise linear) separability of an RNN's hidden-state vectors into semantically different classes. We continue the analysis by computing clusters over the hidden-state vector space with multiple state-of-the-art unsupervised clustering approaches. We formally analyze the accuracy of computed clustering functions and the validity of the clustering hypothesis by determining whether clusters group semantically similar vectors to the same state in the ground-truth model. Our evaluation supports the validity of the clustering hypothesis in the majority of examined cases. We observed that the hidden-state vectors of well-trained RNNs are separable, and that the unsupervised clustering techniques succeed in finding clusters of similar state vectors.

READ FULL TEXT
research
09/10/2020

On Computability, Learnability and Extractability of Finite State Machines from Recurrent Neural Networks

This work aims at shedding some light on connections between finite stat...
research
02/27/2019

Representing Formal Languages: A Comparison Between Finite Automata and Recurrent Neural Networks

We investigate the internal representations that a recurrent neural netw...
research
01/28/2022

Extracting Finite Automata from RNNs Using State Merging

One way to interpret the behavior of a blackbox recurrent neural network...
research
05/09/2021

DocSCAN: Unsupervised Text Classification via Learning from Neighbors

We introduce DocSCAN, a completely unsupervised text classification appr...
research
06/14/2019

On the Computational Power of RNNs

Recent neural network architectures such as the basic recurrent neural n...
research
07/04/2018

Connecting Weighted Automata and Recurrent Neural Networks through Spectral Learning

In this paper, we unravel a fundamental connection between weighted fini...
research
11/23/2016

Emergent Predication Structure in Hidden State Vectors of Neural Readers

A significant number of neural architectures for reading comprehension h...

Please sign up or login with your details

Forgot password? Click here to reset