Implicit N-grams Induced by Recurrence

05/05/2022
by   Xiaobing Sun, et al.
0

Although self-attention based models such as Transformers have achieved remarkable successes on natural language processing (NLP) tasks, recent studies reveal that they have limitations on modeling sequential transformations (Hahn, 2020), which may prompt re-examinations of recurrent neural networks (RNNs) that demonstrated impressive results on handling sequential data. Despite many prior attempts to interpret RNNs, their internal mechanisms have not been fully understood, and the question on how exactly they capture sequential features remains largely unclear. In this work, we present a study that shows there actually exist some explainable components that reside within the hidden states, which are reminiscent of the classical n-grams features. We evaluated such extracted explainable features from trained RNNs on downstream sentiment analysis tasks and found they could be used to model interesting linguistic phenomena such as negation and intensification. Furthermore, we examined the efficacy of using such n-gram components alone as encoders on tasks such as sentiment analysis and language modeling, revealing they could be playing important roles in contributing to the overall performance of RNNs. We hope our findings could add interpretability to RNN architectures, and also provide inspirations for proposing new architectures for sequential data.

READ FULL TEXT
research
12/16/2021

Explainable Natural Language Processing with Matrix Product States

Despite empirical successes of recurrent neural networks (RNNs) in natur...
research
10/30/2017

Understanding Hidden Memories of Recurrent Neural Networks

Recurrent neural networks (RNNs) have been successfully applied to vario...
research
06/25/2019

Reverse engineering recurrent networks for sentiment classification reveals line attractor dynamics

Recurrent neural networks (RNNs) are a widely used tool for modeling seq...
research
07/13/2023

Convolutional Neural Networks for Sentiment Analysis on Weibo Data: A Natural Language Processing Approach

This study addressed the complex task of sentiment analysis on a dataset...
research
03/02/2023

DeepSeer: Interactive RNN Explanation and Debugging via State Abstraction

Recurrent Neural Networks (RNNs) have been widely used in Natural Langua...
research
06/28/2023

Taqyim: Evaluating Arabic NLP Tasks Using ChatGPT Models

Large language models (LLMs) have demonstrated impressive performance on...
research
12/13/2020

MEME: Generating RNN Model Explanations via Model Extraction

Recurrent Neural Networks (RNNs) have achieved remarkable performance on...

Please sign up or login with your details

Forgot password? Click here to reset