Better and Simpler Error Analysis of the Sinkhorn-Knopp Algorithm for Matrix Scaling

01/09/2018
by   Deeparnab Chakrabarty, et al.
0

Given a non-negative n × m real matrix A, the matrix scaling problem is to determine if it is possible to scale the rows and columns so that each row and each column sums to a specified target value for it. This problem arises in many algorithmic applications, perhaps most notably as a preconditioning step in solving a linear system of equations. One of the most natural and by now classical approach to matrix scaling is the Sinkhorn-Knopp algorithm (also known as the RAS method) where one alternately scales either all rows or all columns to meet the target values. In addition to being extremely simple and natural, another appeal of this procedure is that it easily lends itself to parallelization. A central question is to understand the rate of convergence of the Sinkhorn-Knopp algorithm. In this paper, we present an elementary convergence analysis for the Sinkhorn-Knopp algorithm that improves upon the previous best bound. In a nutshell, our approach is to show a simple bound on the number of iterations needed so that the KL-divergence between the current row-sums and the target row-sums drops below a specified threshold δ, and then connect the KL-divergence with ℓ_1 and ℓ_2 distances. For ℓ_1, we can use Pinsker's inequality. For ℓ_2, we develop a strengthening of Pinsker's inequality, called (KL vs ℓ_1/ℓ_2) in the paper, which lower bounds the KL-divergence by a combination of ℓ_1 and ℓ_2 distance. This inequality may be of independent interest. The idea of studying Sinkhorn-Knopp convergence via KL-divergence is not new and has indeed been previously explored. Our contribution is an elementary, self-contained presentation of this approach and an interesting new inequality that yields a significantly stronger convergence guarantee for the extensively studied ℓ_2-error.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/02/2022

Convergence in KL Divergence of the Inexact Langevin Algorithm with Application to Score-based Generative Models

We study the Inexact Langevin Algorithm (ILA) for sampling using estimat...
research
02/15/2022

A short note on an inequality between KL and TV

The goal of this short note is to discuss the relation between Kullback–...
research
11/04/2019

Proximal Langevin Algorithm: Rapid Convergence Under Isoperimetry

We study the Proximal Langevin Algorithm (PLA) for sampling from a proba...
research
12/11/2020

Scaling positive random matrices: concentration and asymptotic convergence

It is well known that any positive matrix can be scaled to have prescrib...
research
09/30/2021

Improved quantum lower and upper bounds for matrix scaling

Matrix scaling is a simple to state, yet widely applicable linear-algebr...
research
04/19/2021

Linear-time uniform generation of random sparse contingency tables with specified marginals

We give an algorithm that generates a uniformly random contingency table...
research
03/19/2017

Bernoulli Rank-1 Bandits for Click Feedback

The probability that a user will click a search result depends both on i...

Please sign up or login with your details

Forgot password? Click here to reset