Strategyproof Learning: Building Trustworthy User-Generated Datasets

06/04/2021
by   Sadegh Farhadkhani, et al.
0

Today's large-scale machine learning algorithms harness massive amounts of user-generated data to train large models. However, especially in the context of content recommendation with enormous social, economical and political incentives to promote specific views, products or ideologies, strategic users might be tempted to fabricate or mislabel data in order to bias algorithms in their favor. Unfortunately, today's learning schemes strongly incentivize such strategic data misreporting. This is a major concern, as it endangers the trustworthiness of the entire training datasets, and questions the safety of any algorithm trained on such datasets. In this paper, we show that, perhaps surprisingly, incentivizing data misreporting is not a fatality. We propose the first personalized collaborative learning framework, Licchavi, with provable strategyproofness guarantees through a careful design of the underlying loss function. Interestingly, we also prove that Licchavi is Byzantine resilient: it tolerates a minority of users that provide arbitrary data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2023

Recommending to Strategic Users

Recommendation systems are pervasive in the digital economy. An importan...
research
02/18/2021

Random Walks with Erasure: Diversifying Personalized Recommendations on Social and Information Networks

Most existing personalization systems promote items that match a user's ...
research
05/31/2022

Strategic Classification with Graph Neural Networks

Strategic classification studies learning in settings where users can mo...
research
02/08/2023

Performative Recommendation: Diversifying Content via Strategic Incentives

The primary goal in recommendation is to suggest relevant content to use...
research
06/11/2021

A Large-Scale Rich Context Query and Recommendation Dataset in Online Knowledge-Sharing

Data plays a vital role in machine learning studies. In the research of ...
research
09/30/2022

SoK: On the Impossible Security of Very Large Foundation Models

Large machine learning models, or so-called foundation models, aim to se...
research
08/05/2021

Recommending Insurance products by using Users' Sentiments

In today's tech-savvy world every industry is trying to formulate method...

Please sign up or login with your details

Forgot password? Click here to reset