Dimension Reduction Approach for Interpretability of Sequence to Sequence Recurrent Neural Networks

05/29/2019
by   Kun Su, et al.
0

Encoder-decoder recurrent neural network models (Seq2Seq) have achieved great success in ubiquitous areas of computation and applications. It was shown to be successful in modeling data with both temporal and spatial dependencies for translation or prediction tasks. In this study, we propose a dimension reduction approach to visualize and interpret the representation of the data by these models. We propose to view the hidden states of the encoder and the decoder as spatio-temporal snapshots of network dynamics and to apply proper orthogonal decomposition to their concatenation to compute a low-dimensional embedding for hidden state dynamics. Projection of the decoder states onto such interpretable embedding space shows that Seq2Seq training to predict sequences using gradient-descent back propagation effectively performs dimension reduction consisting of only a small percentage of dimensions of the network's hidden units. Furthermore, sequences are being clustered into well separable clusters in the low dimensional space each of which corresponds to a different type of dynamics. The projection methodology also clarifies the roles of the encoder and the decoder components of the network. We show that the projection of encoder hidden states onto the low dimensional space provides an initializing trajectory directing the sequence to the cluster which corresponds to that particular type of distinct dynamics and the projection of the decoder hidden states constitutes the embedded cluster attractor. Inspection of the low dimensional space and the projections onto it during training shows that the estimation of clusters separability in the embedding can be utilized to estimate the optimality of model training. We test and demonstrate our proposed interpretability methodology on synthetic examples (dynamics on a circle and an ellipse) and on 3D human body movement data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2021

Understanding How Encoder-Decoder Architectures Attend

Encoder-decoder networks with attention have proven to be a powerful way...
research
11/17/2022

Interpretable Dimensionality Reduction by Feature Preserving Manifold Approximation and Projection

Nonlinear dimensionality reduction lacks interpretability due to the abs...
research
11/27/2019

PREDICT CLUSTER: Unsupervised Skeleton Based Action Recognition

We propose a novel system for unsupervised skeleton-based action recogni...
research
07/25/2018

Multi-view Reconstructive Preserving Embedding for Dimension Reduction

With the development of feature extraction technique, one sample always ...
research
08/18/2023

Latent State Models of Training Dynamics

The impact of randomness on model training is poorly understood. How do ...
research
02/07/2021

Noise Reduction in X-ray Photon Correlation Spectroscopy with Convolutional Neural Networks Encoder-Decoder Models

Like other experimental techniques, X-ray Photon Correlation Spectroscop...
research
08/21/2020

Model reduction in Smoluchowski-type equations

In this paper we utilize the Proper Orthogonal Decomposition (POD) metho...

Please sign up or login with your details

Forgot password? Click here to reset