Towards robust vision by multi-task learning on monkey visual cortex

07/29/2021
by   Shahd Safarani, et al.
4

Deep neural networks set the state-of-the-art across many tasks in computer vision, but their generalization ability to image distortions is surprisingly fragile. In contrast, the mammalian visual system is robust to a wide range of perturbations. Recent work suggests that this generalization ability can be explained by useful inductive biases encoded in the representations of visual stimuli throughout the visual cortex. Here, we successfully leveraged these inductive biases with a multi-task learning approach: we jointly trained a deep network to perform image classification and to predict neural activity in macaque primary visual cortex (V1). We measured the out-of-distribution generalization abilities of our network by testing its robustness to image distortions. We found that co-training on monkey V1 data leads to increased robustness despite the absence of those distortions during training. Additionally, we showed that our network's robustness is very close to that of an Oracle network where parts of the architecture are directly trained on noisy images. Our results also demonstrated that the network's representations become more brain-like as their robustness improves. Using a novel constrained reconstruction analysis, we investigated what makes our brain-regularized network more robust. We found that our co-trained network is more sensitive to content than noise when compared to a Baseline network that we trained for image classification alone. Using DeepGaze-predicted saliency maps for ImageNet images, we found that our monkey co-trained network tends to be more sensitive to salient regions in a scene, reminiscent of existing theories on the role of V1 in the detection of object borders and bottom-up saliency. Overall, our work expands the promising research avenue of transferring inductive biases from the brain, and provides a novel analysis of the effects of our transfer.

READ FULL TEXT

page 7

page 8

page 14

page 15

page 16

research
05/28/2021

Efficient and robust multi-task learning in the brain with modular task primitives

In a real-world setting biological agents do not have infinite resources...
research
11/11/2019

Learning From Brains How to Regularize Machines

Despite impressive performance on numerous visual tasks, Convolutional N...
research
10/25/2018

Decoding Brain Representations by Multimodal Learning of Neural Activity and Visual Features

This paper tackles the problem of learning brain-visual representations ...
research
02/15/2022

Multi-task UNet: Jointly Boosting Saliency Prediction and Disease Classification on Chest X-ray Images

Human visual attention has recently shown its distinct capability in boo...
research
03/13/2023

Evaluating Visual Number Discrimination in Deep Neural Networks

The ability to discriminate between large and small quantities is a core...
research
04/02/2017

Aligned Image-Word Representations Improve Inductive Transfer Across Vision-Language Tasks

An important goal of computer vision is to build systems that learn visu...
research
06/27/2020

On the generalization of learning-based 3D reconstruction

State-of-the-art learning-based monocular 3D reconstruction methods lear...

Please sign up or login with your details

Forgot password? Click here to reset