Stabilizing Equilibrium Models by Jacobian Regularization

06/28/2021
by   Shaojie Bai, et al.
0

Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer. These models have been shown to achieve performance competitive with the state-of-the-art deep networks while using significantly less memory. Yet they are also slower, brittle to architectural choices, and introduce potential instability to the model. In this paper, we propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models. We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains (e.g., WikiText-103 language modeling and ImageNet classification). Using this method, we demonstrate, for the first time, an implicit-depth model that runs with approximately the same speed and level of performance as popular conventional deep networks such as ResNet-101, while still maintaining the constant memory footprint and architectural simplicity of DEQs. Code is available at https://github.com/locuslab/deq .

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/03/2019

Deep Equilibrium Models

We present a new approach to modeling sequential data: the deep equilibr...
research
05/27/2021

Optimization Induced Equilibrium Networks

Implicit equilibrium models, i.e., deep neural networks (DNNs) defined b...
research
04/18/2022

Deep Equilibrium Optical Flow Estimation

Many recent state-of-the-art (SOTA) optical flow models use finite-step ...
research
06/15/2020

Monotone operator equilibrium networks

Implicit-depth models such as Deep Equilibrium Networks have recently be...
research
11/25/2021

Joint inference and input optimization in equilibrium networks

Many tasks in deep learning involve optimizing over the inputs to a netw...
research
10/14/2021

Non-deep Networks

Depth is the hallmark of deep neural networks. But more depth means more...
research
07/15/2022

Stable Invariant Models via Koopman Spectra

Weight-tied models have attracted attention in the modern development of...

Please sign up or login with your details

Forgot password? Click here to reset