Cherry Hypothesis: Identifying the Cherry on the Cake for Dynamic Networks

11/10/2022
by   Shwai He, et al.
0

Dynamic networks have been extensively explored as they can considerably improve the model's representation power with acceptable computational cost. The common practice in implementing dynamic networks is to convert given static layers into fully dynamic ones where all parameters are dynamic and vary with the input. Recent studies empirically show the trend that the more dynamic layers contribute to ever-increasing performance. However, such a fully dynamic setting 1) may cause redundant parameters and high deployment costs, limiting the applicability of dynamic networks to a broader range of tasks and models, and more importantly, 2) contradicts the previous discovery in the human brain that when human brains process an attention-demanding task, only partial neurons in the task-specific areas are activated by the input, while the rest neurons leave in a baseline state. Critically, there is no effort to understand and resolve the above contradictory finding, leaving the primal question – to make the computational parameters fully dynamic or not? – unanswered. The main contributions of our work are challenging the basic commonsense in dynamic networks, and, proposing and validating the cherry hypothesis – A fully dynamic network contains a subset of dynamic parameters that when transforming other dynamic parameters into static ones, can maintain or even exceed the performance of the original network. Technically, we propose a brain-inspired partially dynamic network, namely PAD-Net, to transform the redundant dynamic parameters into static ones. Also, we further design Iterative Mode Partition to partition the dynamic- and static-subnet, which alleviates the redundancy in traditional fully dynamic networks. Our hypothesis and method are comprehensively supported by large-scale experiments with typical advanced dynamic methods.

READ FULL TEXT
research
09/05/2023

Dynamic Brain Transformer with Multi-level Attention for Functional Brain Network Analysis

Recent neuroimaging studies have highlighted the importance of network-c...
research
11/16/2019

S2DNAS: Transforming Static CNN Model for Dynamic Inference via Neural Architecture Search

Recently, dynamic inference has emerged as a promising way to reduce the...
research
07/29/2020

Fully Dynamic Inference with Deep Neural Networks

Modern deep neural networks are powerful and widely applicable models th...
research
05/27/2021

Learning Dynamic Graph Representation of Brain Connectome with Spatio-Temporal Attention

Functional connectivity (FC) between regions of the brain can be assesse...
research
01/10/2023

Cross-Model Comparative Loss for Enhancing Neuronal Utility in Language Understanding

Current natural language understanding (NLU) models have been continuous...
research
01/14/2020

Block-wise Dynamic Sparseness

Neural networks have achieved state of the art performance across a wide...

Please sign up or login with your details

Forgot password? Click here to reset