Learning Transductions and Alignments with RNN Seq2seq Models

03/13/2023
by   Zhengxiang Wang, et al.
0

The paper studies the capabilities of Recurrent-Neural-Network sequence to sequence (RNN seq2seq) models in learning four string-to-string transduction tasks: identity, reversal, total reduplication, and input-specified reduplication. These transductions are traditionally well studied under finite state transducers and attributed with varying complexity. We find that RNN seq2seq models are only able to approximate a mapping that fits the training or in-distribution data. Attention helps significantly, but does not solve the out-of-distribution generalization limitation. Task complexity and RNN variants also play a role in the results. Our results are best understood in terms of the complexity hierarchy of formal languages as opposed to that of string transductions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/09/2017

Handwritten digit string recognition by combination of residual network and RNN-CTC

Recurrent neural network (RNN) and connectionist temporal classification...
research
04/18/2020

A Formal Hierarchy of RNN Architectures

We develop a formal hierarchy of the expressive capacity of RNN architec...
research
02/09/2018

Recurrent Neural Network-Based Semantic Variational Autoencoder for Sequence-to-Sequence Learning

Sequence-to-sequence (Seq2seq) models have played an import role in the ...
research
07/29/2019

RNNbow: Visualizing Learning via Backpropagation Gradients in Recurrent Neural Networks

We present RNNbow, an interactive tool for visualizing the gradient flow...
research
01/25/2016

Survey on the attention based RNN model and its applications in computer vision

The recurrent neural networks (RNN) can be used to solve the sequence to...
research
11/25/2018

Deep RNN Framework for Visual Sequential Applications

Extracting temporal and representation features efficiently plays a pivo...
research
04/01/2021

Do RNN States Encode Abstract Phonological Processes?

Sequence-to-sequence models have delivered impressive results in word fo...

Please sign up or login with your details

Forgot password? Click here to reset