Adaptive Sample-Efficient Blackbox Optimization via ES-active Subspaces

03/07/2019
by   Krzysztof Choromanski, et al.
12

We present a new algorithm ASEBO for conducting optimization of high-dimensional blackbox functions. ASEBO adapts to the geometry of the function and learns optimal sets of sensing directions, which are used to probe it, on-the-fly. It addresses the exploration-exploitation trade-off of blackbox optimization, where each single function query is expensive, by continuously learning the bias of the lower-dimensional model used to approximate gradients of smoothings of the function with compressed sensing and contextual bandits methods. To obtain this model, it uses techniques from the emerging theory of active subspaces in the novel ES blackbox optimization context. As a result, ASEBO learns the dynamically changing intrinsic dimensionality of the gradient space and adapts to the hardness of different stages of the optimization without external supervision. Consequently, it leads to more sample-efficient blackbox optimization than state-of-the-art algorithms. We provide rigorous theoretical justification of the effectiveness of our method. We also empirically evaluate it on the set of reinforcement learning policy optimization tasks as well as functions from the recently open-sourced Nevergrad library, demonstrating that it consistently learns optimal inputs with fewer queries to a blackbox function than other methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/13/2019

ISL: Optimal Policy Learning With Optimal Exploration-Exploitation Trade-Off

Traditionally, off-policy learning algorithms (such as Q-learning) and e...
research
02/07/2020

Ready Policy One: World Building Through Active Learning

Model-Based Reinforcement Learning (MBRL) offers a promising direction f...
research
06/12/2020

A supervised learning approach involving active subspaces for an efficient genetic algorithm in high-dimensional optimization problems

In this work, we present an extension of the genetic algorithm (GA) whic...
research
11/20/2019

Bayesian Curiosity for Efficient Exploration in Reinforcement Learning

Balancing exploration and exploitation is a fundamental part of reinforc...
research
05/31/2023

Representation-Driven Reinforcement Learning

We present a representation-driven framework for reinforcement learning....
research
07/09/2017

Exploiting Active Subspaces in Global Optimization: How Complex is your Problem?

When applying optimization method to a real-world problem, the possessio...

Please sign up or login with your details

Forgot password? Click here to reset