Rover Descent: Learning to optimize by learning to navigate on prototypical loss surfaces

01/22/2018
by   Louis Faury, et al.
0

Learning to optimize - the idea that we can learn from data algorithms that optimize a numerical criterion - has recently been at the heart of a growing number of research efforts. One of the most challenging issues within this approach is to learn a policy that is able to optimize over classes of functions that are fairly different from the ones that it was trained on. We propose a novel way of framing learning to optimize as a problem of learning a good navigation policy on a partially observable loss surface. To this end, we develop Rover Descent, a solution that allows us to learn a fairly broad optimization policy from training on a small set of prototypical two-dimensional surfaces that encompasses the classically hard cases such as valleys, plateaus, cliffs and saddles and by using strictly zero-order information. We show that, without having access to gradient or curvature information, we achieve state-of-the-art convergence speed on optimization problems not presented at training time such as the Rosenbrock function and other hard cases in two dimensions. We extend our framework to optimize over high dimensional landscapes, while still handling only two-dimensional local landscape information and show good preliminary results.

READ FULL TEXT

page 4

page 15

page 16

page 17

research
09/26/2019

GradVis: Visualization and Second Order Analysis of Optimization Surfaces during the Training of Deep Neural Networks

Current training methods for deep neural networks boil down to very high...
research
07/29/2013

A new approach in dynamic traveling salesman problem: a hybrid of ant colony optimization and descending gradient

Nowadays swarm intelligence-based algorithms are being used widely to op...
research
03/24/2018

Gradient descent in Gaussian random fields as a toy model for high-dimensional optimisation in deep learning

In this paper we model the loss function of high-dimensional optimizatio...
research
12/04/2020

Divide and Learn: A Divide and Conquer Approach for Predict+Optimize

The predict+optimize problem combines machine learning ofproblem coeffic...
research
06/18/2020

An adaptive stochastic gradient-free approach for high-dimensional blackbox optimization

In this work, we propose a novel adaptive stochastic gradient-free (ASGF...
research
07/14/2020

Estimating Barycenters of Measures in High Dimensions

Barycentric averaging is a principled way of summarizing populations of ...
research
01/14/2020

On the Convex Behavior of Deep Neural Networks in Relation to the Layers' Width

The Hessian of neural networks can be decomposed into a sum of two matri...

Please sign up or login with your details

Forgot password? Click here to reset