Communication-Efficient Federated Hypergradient Computation via Aggregated Iterative Differentiation

02/09/2023
by   Peiyao Xiao, et al.
0

Federated bilevel optimization has attracted increasing attention due to emerging machine learning and communication applications. The biggest challenge lies in computing the gradient of the upper-level objective function (i.e., hypergradient) in the federated setting due to the nonlinear and distributed construction of a series of global Hessian matrices. In this paper, we propose a novel communication-efficient federated hypergradient estimator via aggregated iterative differentiation (AggITD). AggITD is simple to implement and significantly reduces the communication cost by conducting the federated hypergradient estimation and the lower-level optimization simultaneously. We show that the proposed AggITD-based algorithm achieves the same sample complexity as existing approximate implicit differentiation (AID)-based approaches with much fewer communication rounds in the presence of data heterogeneity. Our results also shed light on the great advantage of ITD over AID in the federated/distributed hypergradient estimation. This differs from the comparison in the non-distributed bilevel optimization, where ITD is less efficient than AID. Our extensive experiments demonstrate the great effectiveness and communication efficiency of the proposed method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2023

Communication-Efficient Federated Bilevel Optimization with Local and Global Lower Level Problems

Bilevel Optimization has witnessed notable progress recently with new em...
research
11/02/2022

Fast Adaptive Federated Bilevel Optimization

Bilevel optimization is a popular hierarchical model in machine learning...
research
05/30/2023

SimFBO: Towards Simple, Flexible and Communication-efficient Federated Bilevel Learning

Federated bilevel optimization (FBO) has shown great potential recently ...
research
05/18/2023

Q-SHED: Distributed Optimization at the Edge via Hessian Eigenvectors Quantization

Edge networks call for communication efficient (low overhead) and robust...
research
04/15/2023

Gradient-less Federated Gradient Boosting Trees with Learnable Learning Rates

The privacy-sensitive nature of decentralized datasets and the robustnes...
research
06/21/2023

An Efficient Virtual Data Generation Method for Reducing Communication in Federated Learning

Communication overhead is one of the major challenges in Federated Learn...
research
08/08/2023

Federated Zeroth-Order Optimization using Trajectory-Informed Surrogate Gradients

Federated optimization, an emerging paradigm which finds wide real-world...

Please sign up or login with your details

Forgot password? Click here to reset