Optimal rates for zero-order convex optimization: the power of two function evaluations

12/07/2013
by   John C. Duchi, et al.
0

We consider derivative-free algorithms for stochastic and non-stochastic convex optimization problems that use only function values rather than gradients. Focusing on non-asymptotic bounds on convergence rates, we show that if pairs of function values are available, algorithms for d-dimensional optimization that use gradient estimates based on random perturbations suffer a factor of at most √(d) in convergence rate over traditional stochastic gradient methods. We establish such results for both smooth and non-smooth cases, sharpening previous analyses that suggested a worse dimension dependence, and extend our results to the case of multiple (m > 2) evaluations. We complement our algorithmic development with information-theoretic lower bounds on the minimax convergence rate of such problems, establishing the sharpness of our achievable results up to constant (sometimes logarithmic) factors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2020

An improved convergence analysis for decentralized online stochastic non-convex optimization

In this paper, we study decentralized online stochastic non-convex optim...
research
02/23/2022

Mirror Descent Strikes Again: Optimal Stochastic Convex Optimization under Infinite Noise Variance

We study stochastic convex optimization under infinite noise variance. S...
research
09/11/2012

Query Complexity of Derivative-Free Optimization

This paper provides lower bounds on the convergence rate of Derivative F...
research
02/02/2022

Tight Convergence Rate Bounds for Optimization Under Power Law Spectral Conditions

Performance of optimization on quadratic problems sensitively depends on...
research
09/11/2021

Fundamental limits of over-the-air optimization: Are analog schemes optimal?

We consider over-the-air convex optimization on a d-dimensional space wh...
research
03/22/2018

Optimization of Smooth Functions with Noisy Observations: Local Minimax Rates

We consider the problem of global optimization of an unknown non-convex ...
research
03/06/2023

Convergence Rates for Non-Log-Concave Sampling and Log-Partition Estimation

Sampling from Gibbs distributions p(x) ∝exp(-V(x)/ε) and computing their...

Please sign up or login with your details

Forgot password? Click here to reset