High-Dimensional Inference over Networks: Linear Convergence and Statistical Guarantees

01/21/2022
by   Ying Sun, et al.
0

We study sparse linear regression over a network of agents, modeled as an undirected graph and no server node. The estimation of the s-sparse parameter is formulated as a constrained LASSO problem wherein each agent owns a subset of the N total observations. We analyze the convergence rate and statistical guarantees of a distributed projected gradient tracking-based algorithm under high-dimensional scaling, allowing the ambient dimension d to grow with (and possibly exceed) the sample size N. Our theory shows that, under standard notions of restricted strong convexity and smoothness of the loss functions, suitable conditions on the network connectivity and algorithm tuning, the distributed algorithm converges globally at a linear rate to an estimate that is within the centralized statistical precision of the model, O(slog d/N). When slog d/N=o(1), a condition necessary for statistical consistency, an ε-optimal solution is attained after 𝒪(κlog (1/ε)) gradient computations and O (κ/(1-ρ) log (1/ε)) communication rounds, where κ is the restricted condition number of the loss function and ρ measures the network connectivity. The computation cost matches that of the centralized projected gradient algorithm despite having data distributed; whereas the communication rounds reduce as the network connectivity improves. Overall, our study reveals interesting connections between statistical efficiency, network connectivity & topology, and convergence rate in high dimensions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2021

Distributed Sparse Regression via Penalization

We study sparse linear regression over a network of agents, modeled as a...
research
04/25/2011

Fast global convergence of gradient methods for high-dimensional statistical recovery

Many statistical M-estimators are based on convex optimization problems ...
research
10/08/2012

A Fast Distributed Proximal-Gradient Method

We present a distributed proximal-gradient method for optimizing the ave...
research
03/22/2018

SUCAG: Stochastic Unbiased Curvature-aided Gradient Method for Distributed Optimization

We propose and analyze a new stochastic gradient method, which we call S...
research
11/07/2016

Linear Convergence of SVRG in Statistical Estimation

SVRG and its variants are among the state of art optimization algorithms...
research
10/09/2019

Straggler-Agnostic and Communication-Efficient Distributed Primal-Dual Algorithm for High-Dimensional Data Mining

Recently, reducing communication time between machines becomes the main ...
research
11/02/2021

Distributed Sparse Feature Selection in Communication-Restricted Networks

This paper aims to propose and theoretically analyze a new distributed s...

Please sign up or login with your details

Forgot password? Click here to reset