On the Computational Complexity of Linear Discrepancy

07/31/2020
by   Lily Li, et al.
0

Many problems in computer science and applied mathematics require rounding a vector 𝐰 of fractional values lying in the interval [0,1] to a binary vector 𝐱 so that, for a given matrix 𝐀, 𝐀𝐱 is as close to 𝐀𝐰 as possible. For example, this problem arises in LP rounding algorithms used to approximate 𝖭𝖯-hard optimization problems and in the design of uniformly distributed point sets for numerical integration. For a given matrix 𝐀, the worst-case error over all choices of 𝐰 incurred by the best possible rounding is measured by the linear discrepancy of 𝐀, a quantity studied in discrepancy theory, and introduced by Lovasz, Spencer, and Vesztergombi (EJC, 1986). We initiate the study of the computational complexity of linear discrepancy. Our investigation proceeds in two directions: (1) proving hardness results and (2) finding both exact and approximate algorithms to evaluate the linear discrepancy of certain matrices. For (1), we show that linear discrepancy is 𝖭𝖯-hard. Thus we do not expect to find an efficient exact algorithm for the general case. Restricting our attention to matrices with a constant number of rows, we present a poly-time exact algorithm for matrices consisting of a single row and matrices with a constant number of rows and entries of bounded magnitude. We also present an exponential-time approximation algorithm for general matrices, and an algorithm that approximates linear discrepancy to within an exponential factor.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/02/2021

Linear Discrepancy is Π_2-Hard to Approximate

In this note, we prove that the problem of computing the linear discrepa...
research
11/28/2021

Computational Complexity of Normalizing Constants for the Product of Determinantal Point Processes

We consider the product of determinantal point processes (DPPs), a point...
research
02/27/2023

Generator Matrices by Solving Integer Linear Programs

In quasi-Monte Carlo methods, generating high-dimensional low discrepanc...
research
07/03/2019

Linear Size Sparsifier and the Geometry of the Operator Norm Ball

The Matrix Spencer Conjecture asks whether given n symmetric matrices in...
research
02/26/2023

Fast Attention Requires Bounded Entries

In modern machine learning, inner product attention computation is a fun...
research
09/19/2018

Simple, fast and accurate evaluation of the action of the exponential of a rate matrix on a probability vector

Given a time-homogeneous, finite-statespace Markov chain with a rate mat...
research
05/15/2023

Linear-Sized Sparsifiers via Near-Linear Time Discrepancy Theory

Discrepancy theory provides powerful tools for producing higher-quality ...

Please sign up or login with your details

Forgot password? Click here to reset