Anonymous Bandits for Multi-User Systems

10/21/2022
by   Hossein Esfandiari, et al.
0

In this work, we present and study a new framework for online learning in systems with multiple users that provide user anonymity. Specifically, we extend the notion of bandits to obey the standard k-anonymity constraint by requiring each observation to be an aggregation of rewards for at least k users. This provides a simple yet effective framework where one can learn a clustering of users in an online fashion without observing any user's individual decision. We initiate the study of anonymous bandits and provide the first sublinear regret algorithms and lower bounds for this setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/13/2011

From Bandits to Experts: On the Value of Side-Observations

We consider an adversarial online learning setting where a decision make...
research
01/05/2021

Sequential Choice Bandits with Feedback for Personalizing users' experience

In this work, we study sequential choice bandits with feedback. We propo...
research
04/13/2020

Power-Constrained Bandits

Contextual bandits often provide simple and effective personalization in...
research
02/25/2019

Improved Algorithm on Online Clustering of Bandits

We generalize the setting of online clustering of bandits by allowing no...
research
03/11/2020

Delay-Adaptive Learning in Generalized Linear Contextual Bandits

In this paper, we consider online learning in generalized linear context...
research
01/30/2022

Coordinated Attacks against Contextual Bandits: Fundamental Limits and Defense Mechanisms

Motivated by online recommendation systems, we propose the problem of fi...
research
10/21/2019

Multi-User MABs with User Dependent Rewards for Uncoordinated Spectrum Access

Multi-user multi-armed bandits have emerged as a good model for uncoordi...

Please sign up or login with your details

Forgot password? Click here to reset