Expectation propagation for the diluted Bayesian classifier

09/20/2020
by   Alfredo Braunstein, et al.
0

Efficient feature selection from high-dimensional datasets is a very important challenge in many data-driven fields of science and engineering. We introduce a statistical mechanics inspired strategy that addresses the problem of sparse feature selection in the context of binary classification by leveraging a computational scheme known as expectation propagation (EP). The algorithm is used in order to train a continuous-weights perceptron learning a classification rule from a set of (possibly partly mislabeled) examples provided by a teacher perceptron with diluted continuous weights. We test the method in the Bayes optimal setting under a variety of conditions and compare it to other state-of-the-art algorithms based on message passing and on expectation maximization approximate inference schemes. Overall, our simulations show that EP is a robust and competitive algorithm in terms of variable selection properties, estimation accuracy and computationally complexity, especially when the student perceptron is trained from correlated patterns that prevent other iterative methods from converging. Furthermore, our numerical tests demonstrate that the algorithm is capable of learning online the unknown values of prior parameters, such as the dilution level of the weights of the teacher perceptron and the fraction of mislabeled examples, quite accurately. This is achieved by means of a simple maximum likelihood strategy that consists in minimizing the free energy associated with the EP algorithm.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/05/2014

Binary Linear Classification and Feature Selection via Generalized Approximate Message Passing

For the problem of binary linear classification and feature selection, w...
research
03/22/2022

Learning curves for the multi-class teacher-student perceptron

One of the most classical results in high-dimensional learning theory pr...
research
09/15/2015

Sparse Multinomial Logistic Regression via Approximate Message Passing

For the problem of multi-class linear classification and feature selecti...
research
02/21/2019

Active online learning in the binary perceptron problem

The binary perceptron is the simplest artificial neural network formed b...
research
03/08/2017

A GAMP Based Low Complexity Sparse Bayesian Learning Algorithm

In this paper, we present an algorithm for the sparse signal recovery pr...
research
11/22/2017

PULasso: High-dimensional variable selection with presence-only data

In various real-world problems, we are presented with positive and unlab...
research
08/07/2023

The Copycat Perceptron: Smashing Barriers Through Collective Learning

We characterize the equilibrium properties of a model of y coupled binar...

Please sign up or login with your details

Forgot password? Click here to reset