Robust Learning from Untrusted Sources

01/29/2019
by   Nikola Konstantinov, et al.
0

Modern machine learning methods often require more data for training than a single expert can provide. Therefore, it has become a standard procedure to collect data from external sources, e.g. via crowdsourcing. Unfortunately, the quality of these sources is not always guaranteed. As additional complications, the data might be stored in a distributed way, or might even have to remain private. In this work, we address the question of how to learn robustly in such scenarios. Studying the problem through the lens of statistical learning theory, we derive a procedure that allows for learning from all available sources, yet automatically suppresses irrelevant or corrupted data. We show by extensive experiments that our method provides significant improvements over alternative approaches from robust statistics and distributed optimization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2022

Gray Learning from Non-IID Data with Out-of-distribution Samples

The quality of the training data annotated by experts cannot be guarante...
research
06/22/2021

FLEA: Provably Fair Multisource Learning from Unreliable Training Data

Fairness-aware learning aims at constructing classifiers that not only m...
research
03/23/2019

Distributed Lossy Image Compression with Recurrent Networks

We propose a new architecture for distributed image compression from a g...
research
08/28/2021

A robust fusion-extraction procedure with summary statistics in the presence of biased sources

Information from various data sources is increasingly available nowadays...
research
09/19/2023

FRAMU: Attention-based Machine Unlearning using Federated Reinforcement Learning

Machine Unlearning is an emerging field that addresses data privacy issu...
research
12/06/2022

Measuring Intangible Assets Using Parametric and Machine Learning Approaches

Intangible capital as the result of digitalization and globalization has...
research
02/18/2021

Robust and Differentially Private Mean Estimation

Differential privacy has emerged as a standard requirement in a variety ...

Please sign up or login with your details

Forgot password? Click here to reset