Top-Down Processing: Top-Down Network Combines Back-Propagation with Attention

06/04/2023
by   Roy Abel, et al.
0

Early neural network models relied exclusively on bottom-up processing going from the input signals to higher-level representations. Many recent models also incorporate top-down networks going in the opposite direction. Top-down processing in deep learning models plays two primary roles: learning and directing attention. These two roles are accomplished in current models through distinct mechanisms. While top-down attention is often implemented by extending the model's architecture with additional units that propagate information from high to low levels of the network, learning is typically accomplished by an external learning algorithm such as back-propagation. In the current work, we present an integration of the two functions above, which appear unrelated, using a single unified mechanism. We propose a novel symmetric bottom-up top-down network structure that can integrate standard bottom-up networks with a symmetric top-down counterpart, allowing each network to guide and influence the other. The same top-down network is being used for both learning, via back-propagating feedback signals, and at the same time also for top-down attention, by guiding the bottom-up network to perform a selected task. We show that our method achieves competitive performance on a standard multi-task learning benchmark. Yet, we rely on standard single-task architectures and optimizers, without any task-specific parameters. Additionally, our learning algorithm addresses in a new way some neuroscience issues that arise in biological modeling of learning in the brain.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2018

End-to-End Multi-Task Learning with Attention

In this paper, we propose a novel multi-task learning architecture, whic...
research
02/18/2020

Multi-Task Learning from Videos via Efficient Inter-Frame Attention

Prior work in multi-task learning has mainly focused on predictions on a...
research
04/11/2023

AdaTT: Adaptive Task-to-Task Fusion Network for Multitask Learning in Recommendations

Multi-task learning (MTL) aims at enhancing the performance and efficien...
research
07/24/2020

Reparameterizing Convolutions for Incremental Multi-Task Learning without Task Interference

Multi-task networks are commonly utilized to alleviate the need for a la...
research
11/15/2022

Cross-Stitched Multi-task Dual Recursive Networks for Unified Single Image Deraining and Desnowing

We present the Cross-stitched Multi-task Unified Dual Recursive Network ...
research
03/04/2020

Unsupervised and Interpretable Domain Adaptation to Rapidly Filter Social Web Data for Emergency Services

During the onset of a disaster event, filtering relevant information fro...
research
11/30/2017

Learning to Adapt by Minimizing Discrepancy

We explore whether useful temporal neural generative models can be learn...

Please sign up or login with your details

Forgot password? Click here to reset