Implicit differentiation for fast hyperparameter selection in non-smooth convex learning

05/04/2021
by   Quentin Bertrand, et al.
0

Finding the optimal hyperparameters of a model can be cast as a bilevel optimization problem, typically solved using zero-order techniques. In this work we study first-order methods when the inner optimization problem is convex but non-smooth. We show that the forward-mode differentiation of proximal gradient descent and proximal coordinate descent yield sequences of Jacobians converging toward the exact Jacobian. Using implicit differentiation, we show it is possible to leverage the non-smoothness of the inner problem to speed up the computation. Finally, we provide a bound on the error made on the hypergradient when the inner optimization problem is solved approximately. Results on regression and classification problems reveal computational benefits for hyperparameter optimization, especially when multiple hyperparameters are required.

READ FULL TEXT
research
07/05/2023

Implicit Differentiation for Hyperparameter Tuning the Weighted Graphical Lasso

We provide a framework and algorithm for tuning the hyperparameters of t...
research
02/20/2020

Implicit differentiation of Lasso-type models for hyperparameter optimization

Setting regularization parameters for Lasso-type estimators is notorious...
research
11/09/2022

Approximate backwards differentiation of gradient flow

The gradient flow (GF) is an ODE for which its explicit Euler's discreti...
research
01/11/2023

Analyzing Inexact Hypergradients for Bilevel Learning

Estimating hyperparameters has been a long-standing problem in machine l...
research
10/26/2022

Coordinate Descent for SLOPE

The lasso is the most famous sparse regression and feature selection met...
research
11/29/2021

Amortized Implicit Differentiation for Stochastic Bilevel Optimization

We study a class of algorithms for solving bilevel optimization problems...
research
10/26/2020

Delta-STN: Efficient Bilevel Optimization for Neural Networks using Structured Response Jacobians

Hyperparameter optimization of neural networks can be elegantly formulat...

Please sign up or login with your details

Forgot password? Click here to reset