Differentiable Fixed-Point Iteration Layer

02/07/2020
by   Younahan Jeon, et al.
0

Recently, several studies proposed methods to utilize some restricted classes of optimization problems as layers of deep neural networks. However, these methods are still in their infancy and require special treatments, i.e., analyzing the KKT condition, etc., for deriving the backpropagation formula. Instead, in this paper, we propose a method to utilize fixed-point iteration (FPI), a generalization of many types of numerical algorithms, as a network layer. We show that the derivative of an FPI layer depends only on the fixed point, and then we present a method to calculate it efficiently using another FPI which we call the backward FPI. The proposed method can be easily implemented based on the autograd functionalities in existing deep learning tools. Since FPI covers vast different types of numerical algorithms in machine learning and other fields, it has a lot of potential applications. In the experiments, the differentiable FPI layer is applied to two scenarios, i.e., gradient descent iterations for differentiable optimization problems and FPI with arbitrary neural network modules, of which the results demonstrate the simplicity and the effectiveness.

READ FULL TEXT

page 6

page 7

research
06/29/2017

A Fixed-Point of View on Gradient Methods for Big Data

Interpreting gradient methods as fixed-point iterations, we provide a de...
research
01/10/2020

Chebyshev Inertial Iteration for Accelerating Fixed-Point Iterations

A novel method which is called the Chebyshev inertial iteration for acce...
research
11/25/2021

Joint inference and input optimization in equilibrium networks

Many tasks in deep learning involve optimizing over the inputs to a netw...
research
08/21/2023

Differentiable Frank-Wolfe Optimization Layer

Differentiable optimization has received a significant amount of attenti...
research
07/28/2014

'Almost Sure' Chaotic Properties of Machine Learning Methods

It has been demonstrated earlier that universal computation is 'almost s...
research
12/14/2021

Efficient differentiable quadratic programming layers: an ADMM approach

Recent advances in neural-network architecture allow for seamless integr...
research
04/10/2023

Approximate Primal-Dual Fixed-Point based Langevin Algorithms for Non-smooth Convex Potentials

The Langevin algorithms are frequently used to sample the posterior dist...

Please sign up or login with your details

Forgot password? Click here to reset