Sparse Perturbations for Improved Convergence in Stochastic Zeroth-Order Optimization

06/02/2020
by   Mayumi Ohta, et al.
0

Interest in stochastic zeroth-order (SZO) methods has recently been revived in black-box optimization scenarios such as adversarial black-box attacks to deep neural networks. SZO methods only require the ability to evaluate the objective function at random input points, however, their weakness is the dependency of their convergence speed on the dimensionality of the function to be evaluated. We present a sparse SZO optimization method that reduces this factor to the expected dimensionality of the random perturbation during learning. We give a proof that justifies this reduction for sparse SZO optimization for non-convex functions without making any assumptions on sparsity of objective function or gradient. Furthermore, we present experimental results for neural networks on MNIST and CIFAR that show faster convergence in training loss and test accuracy, and a smaller distance of the gradient approximation to the true gradient in sparse SZO compared to dense SZO.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/21/2020

Zeroth-Order Hybrid Gradient Descent: Towards A Principled Black-Box Optimization Framework

In this work, we focus on the study of stochastic zeroth-order (ZO) opti...
research
03/29/2020

Zeroth-Order Regularized Optimization (ZORO): Approximately Sparse Gradients and Adaptive Sampling

We consider the problem of minimizing a high-dimensional objective funct...
research
05/30/2018

Stochastic Zeroth-order Optimization via Variance Reduction method

Derivative-free optimization has become an important technique used in m...
research
06/12/2018

Sparse Stochastic Zeroth-Order Optimization with an Application to Bandit Structured Prediction

Stochastic zeroth-order (SZO), or gradient-free, optimization allows to ...
research
10/15/2019

ZO-AdaMM: Zeroth-Order Adaptive Momentum Method for Black-Box Optimization

The adaptive momentum method (AdaMM), which uses past gradients to updat...
research
03/25/2020

Zeroth-order Optimization on Riemannian Manifolds

We propose and analyze zeroth-order algorithms for optimization over Rie...
research
07/13/2019

Distributed Black-Box Optimization via Error Correcting Codes

We introduce a novel distributed derivative-free optimization framework ...

Please sign up or login with your details

Forgot password? Click here to reset