DeepAI AI Chat
Log In Sign Up

Asynchronous Upper Confidence Bound Algorithms for Federated Linear Bandits

by   Chuanhao Li, et al.
University of Virginia

Linear contextual bandit is a popular online learning problem. It has been mostly studied in centralized learning settings. With the surging demand of large-scale decentralized model learning, e.g., federated learning, how to retain regret minimization while reducing communication cost becomes an open challenge. In this paper, we study linear contextual bandit in a federated learning setting. We propose a general framework with asynchronous model update and communication for a collection of homogeneous clients and heterogeneous clients, respectively. Rigorous theoretical analysis is provided about the regret and communication cost under this distributed learning framework; and extensive empirical evaluations demonstrate the effectiveness of our solution.


page 1

page 2

page 3

page 4


Communication Efficient Federated Learning for Generalized Linear Bandits

Contextual bandit algorithms have been recently studied under the federa...

Differentially-Private Federated Linear Bandits

The rapid proliferation of decentralized learning systems mandates the n...

An Empirical Evaluation of Federated Contextual Bandit Algorithms

As the adoption of federated learning increases for learning from sensit...

Federated Multi-armed Bandits with Personalization

A general framework of personalized federated multi-armed bandits (PF-MA...

A Communication-Efficient Adaptive Algorithm for Federated Learning under Cumulative Regret

We consider the problem of online stochastic optimization in a distribut...

A Simple and Provably Efficient Algorithm for Asynchronous Federated Contextual Linear Bandits

We study federated contextual linear bandits, where M agents cooperate w...

A Novel Optimized Asynchronous Federated Learning Framework

Federated Learning (FL) since proposed has been applied in many fields, ...