Modeling Structure with Undirected Neural Networks

02/08/2022
by   Tsvetomila Mihaylova, et al.
0

Neural networks are powerful function estimators, leading to their status as a paradigm of choice for modeling structured data. However, unlike other structured representations that emphasize the modularity of the problem – e.g., factor graphs – neural networks are usually monolithic mappings from inputs to outputs, with a fixed computation order. This limitation prevents them from capturing different directions of computation and interaction between the modeled variables. In this paper, we combine the representational strengths of factor graphs and of neural networks, proposing undirected neural networks (UNNs): a flexible framework for specifying computations that can be performed in any order. For particular choices, our proposed models subsume and extend many existing architectures: feed-forward, recurrent, self-attention networks, auto-encoders, and networks with implicit layers. We demonstrate the effectiveness of undirected neural architectures, both unstructured and structured, on a range of tasks: tree-constrained dependency parsing, convolutional image classification, and sequence completion with attention. By varying the computation order, we show how a single UNN can be used both as a classifier and a prototype generator, and how it can fill in missing parts of an input sequence, making them a promising field for further research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/28/2020

Temporal Convolutional Attention-based Network For Sequence Modeling

With the development of feed-forward models, the default model for seque...
research
10/20/2020

Implicit recurrent networks: A novel approach to stationary input processing with recurrent neural networks in deep learning

The brain cortex, which processes visual, auditory and sensory data in t...
research
05/28/2020

When Can Self-Attention Be Replaced by Feed Forward Layers?

Recently, self-attention models such as Transformers have given competit...
research
08/01/2017

Natural Language Processing with Small Feed-Forward Networks

We show that small and shallow feed-forward neural networks can achieve ...
research
04/19/2023

Learning Hierarchically-Structured Concepts II: Overlapping Concepts, and Networks With Feedback

We continue our study from Lynch and Mallmann-Trenn (Neural Networks, 20...
research
08/24/2021

Auto-Parsing Network for Image Captioning and Visual Question Answering

We propose an Auto-Parsing Network (APN) to discover and exploit the inp...

Please sign up or login with your details

Forgot password? Click here to reset