A scaled Bregman theorem with applications

07/01/2016
by   Richard Nock, et al.
0

Bregman divergences play a central role in the design and analysis of a range of machine learning algorithms. This paper explores the use of Bregman divergences to establish reductions between such algorithms and their analyses. We present a new scaled isodistortion theorem involving Bregman divergences (scaled Bregman theorem for short) which shows that certain "Bregman distortions'" (employing a potentially non-convex generator) may be exactly re-written as a scaled Bregman divergence computed over transformed data. Admissible distortions include geodesic distances on curved manifolds and projections or gauge-normalisation, while admissible data include scalars, vectors and matrices. Our theorem allows one to leverage to the wealth and convenience of Bregman divergences when analysing algorithms relying on the aforementioned Bregman distortions. We illustrate this with three novel applications of our theorem: a reduction from multi-class density ratio to class-probability estimation, a new adaptive projection free yet norm-enforcing dual norm mirror descent algorithm, and a reduction from clustering on flat manifolds to clustering on curved manifolds. Experiments on each of these domains validate the analyses and suggest that the scaled Bregman theorem might be a worthy addition to the popular handful of Bregman divergence properties that have been pervasive in machine learning.

READ FULL TEXT
research
12/18/2014

Stochastic Descent Analysis of Representation Learning Algorithms

Although stochastic approximation learning methods have been widely used...
research
09/01/2019

Stability of the Cut Locus and a Central Limit Theorem for Fréchet Means of Riemannian Manifolds

We obtain a Central Limit Theorem for closed Riemannian manifolds, clari...
research
06/08/2020

Learning disconnected manifolds: a no GANs land

Typical architectures of Generative AdversarialNetworks make use of a un...
research
05/18/2020

Stable and consistent density-based clustering

We present a consistent approach to density-based clustering, which sati...
research
02/13/2012

Sparse Matrix Inversion with Scaled Lasso

We propose a new method of learning a sparse nonnegative-definite target...
research
06/01/2022

Stochastic Gradient Methods with Preconditioned Updates

This work considers non-convex finite sum minimization. There are a numb...

Please sign up or login with your details

Forgot password? Click here to reset