Convex optimization over a probability simplex

05/15/2023
by   James Chok, et al.
0

We propose a new iteration scheme, the Cauchy-Simplex, to optimize convex problems over the probability simplex {w∈ℝ^n | ∑_i w_i=1and w_i≥0}. Other works have taken steps to enforce positivity or unit normalization automatically but never simultaneously within a unified setting. This paper presents a natural framework for manifestly requiring the probability condition. Specifically, we map the simplex to the positive quadrant of a unit sphere, envisage gradient descent in latent variables, and map the result back in a way that only depends on the simplex variable. Moreover, proving rigorous convergence results in this formulation leads inherently to tools from information theory (e.g. cross entropy and KL divergence). Each iteration of the Cauchy-Simplex consists of simple operations, making it well-suited for high-dimensional problems. We prove that it has a convergence rate of O(1/T) for convex functions, and numerical experiments of projection onto convex hulls show faster convergence than similar algorithms. Finally, we apply our algorithm to online learning problems and prove the convergence of the average regret for (1) Prediction with expert advice and (2) Universal Portfolios.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2015

Random Multi-Constraint Projection: Stochastic Gradient Methods for Convex Optimization with Many Constraints

Consider convex optimization problems subject to a large number of const...
research
07/12/2023

Provably Faster Gradient Descent via Long Steps

This work establishes provably faster convergence rates for gradient des...
research
02/17/2020

Last iterate convergence in no-regret learning: constrained min-max optimization for convex-concave landscapes

In a recent series of papers it has been established that variants of Gr...
research
02/27/2022

Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization

We consider linear prediction with a convex Lipschitz loss, or more gene...
research
02/03/2023

Pseudonorm Approachability and Applications to Regret Minimization

Blackwell's celebrated approachability theory provides a general framewo...
research
05/12/2009

Experiment Study of Entropy Convergence of Ant Colony Optimization

Ant colony optimization (ACO) has been applied to the field of combinato...
research
09/06/2015

Deep Online Convex Optimization by Putting Forecaster to Sleep

Methods from convex optimization such as accelerated gradient descent ar...

Please sign up or login with your details

Forgot password? Click here to reset