Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box Optimization Framework

12/21/2020
by   Pranay Sharma, et al.
17

In this work, we focus on the study of stochastic zeroth-order (ZO) optimization which does not require first-order gradient information and uses only function evaluations. The problem of ZO optimization has emerged in many recent machine learning applications, where the gradient of the objective function is either unavailable or difficult to compute. In such cases, we can approximate the full gradients or stochastic gradients through function value based gradient estimates. Here, we propose a novel hybrid gradient estimator (HGE), which takes advantage of the query-efficiency of random gradient estimates as well as the variance-reduction of coordinate-wise gradient estimates. We show that with a graceful design in coordinate importance sampling, the proposed HGE-based ZO optimization method is efficient both in terms of iteration complexity as well as function query cost. We provide a thorough theoretical analysis of the convergence of our proposed method for non-convex, convex, and strongly-convex optimization. We show that the convergence rate that we derive generalizes the results for some prominent existing methods in the nonconvex case, and matches the optimal result in the convex case. We also corroborate the theory with a real-world black-box attack generation application to demonstrate the empirical advantage of our method over state-of-the-art ZO optimization approaches.

READ FULL TEXT
research
09/30/2019

Min-Max Optimization without Gradients: Convergence and Applications to Adversarial ML

In this paper, we study the problem of constrained robust (min-max) opti...
research
06/02/2020

Sparse Perturbations for Improved Convergence in Stochastic Zeroth-Order Optimization

Interest in stochastic zeroth-order (SZO) methods has recently been revi...
research
10/11/2022

Zeroth-Order Hard-Thresholding: Gradient Error vs. Expansivity

ℓ_0 constrained optimization is prevalent in machine learning, particula...
research
03/29/2020

Zeroth-Order Regularized Optimization (ZORO): Approximately Sparse Gradients and Adaptive Sampling

We consider the problem of minimizing a high-dimensional objective funct...
research
05/30/2018

Stochastic Zeroth-order Optimization via Variance Reduction method

Derivative-free optimization has become an important technique used in m...
research
07/17/2016

Global Continuous Optimization with Error Bound and Fast Convergence

This paper considers global optimization with a black-box unknown object...
research
05/25/2018

Zeroth-Order Stochastic Variance Reduction for Nonconvex Optimization

As application demands for zeroth-order (gradient-free) optimization acc...

Please sign up or login with your details

Forgot password? Click here to reset