Non-Asymptotic Performance of Social Machine Learning Under Limited Data

06/15/2023
by   Ping Hu, et al.
0

This paper studies the probability of error associated with the social machine learning framework, which involves an independent training phase followed by a cooperative decision-making phase over a graph. This framework addresses the problem of classifying a stream of unlabeled data in a distributed manner. We consider two kinds of classification tasks with limited observations in the prediction phase, namely, the statistical classification task and the single-sample classification task. For each task, we describe the distributed learning rule and analyze the probability of error accordingly. To do so, we first introduce a stronger consistent training condition that involves the margin distributions generated by the trained classifiers. Based on this condition, we derive an upper bound on the probability of error for both tasks, which depends on the statistical properties of the data and the combination policy used to combine the distributed classifiers. For the statistical classification problem, we employ the geometric social learning rule and conduct a non-asymptotic performance analysis. An exponential decay of the probability of error with respect to the number of unlabeled samples is observed in the upper bound. For the single-sample classification task, a distributed learning rule that functions as an ensemble classifier is constructed. An upper bound on the probability of error of this ensemble classifier is established.

READ FULL TEXT

page 1

page 11

page 13

research
12/17/2021

Learning from Heterogeneous Data Based on Social Interactions over Graphs

This work proposes a decentralized architecture, where individual agents...
research
10/23/2020

Network Classifiers Based on Social Learning

This work proposes a new way of combining independently trained classifi...
research
07/29/2009

Cooperative Training for Attribute-Distributed Data: Trade-off Between Data Transmission and Performance

This paper introduces a modeling framework for distributed regression wi...
research
06/29/2016

Tighter bounds lead to improved classifiers

The standard approach to supervised classification involves the minimiza...
research
06/02/2021

Statistical optimality conditions for compressive ensembles

We present a framework for the theoretical analysis of ensembles of low-...
research
03/05/2015

Optimally Combining Classifiers Using Unlabeled Data

We develop a worst-case analysis of aggregation of classifier ensembles ...

Please sign up or login with your details

Forgot password? Click here to reset