Regret Bounds for Non-decomposable Metrics with Missing Labels

06/07/2016
by   Prateek Jain, et al.
0

We consider the problem of recommending relevant labels (items) for a given data point (user). In particular, we are interested in the practically important setting where the evaluation is with respect to non-decomposable (over labels) performance metrics like the F_1 measure, and the training data has missing labels. To this end, we propose a generic framework that given a performance metric Ψ, can devise a regularized objective function and a threshold such that all the values in the predicted score vector above and only above the threshold are selected to be positive. We show that the regret or generalization error in the given metric Ψ is bounded ultimately by estimation error of certain underlying parameters. In particular, we derive regret bounds under three popular settings: a) collaborative filtering, b) multilabel classification, and c) PU (positive-unlabeled) learning. For each of the above problems, we can obtain precise non-asymptotic regret bound which is small even when a large fraction of labels is missing. Our empirical results on synthetic and benchmark datasets demonstrate that by explicitly modeling for missing labels and optimizing the desired performance metric, our algorithm indeed achieves significantly better performance (like F_1 score) when compared to methods that do not model missing label information carefully.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2017

Regret Bounds and Regimes of Optimality for User-User and Item-Item Collaborative Filtering

We consider an online model for recommendation systems, with each user b...
research
03/25/2021

Prediction in the presence of response-dependent missing labels

In a variety of settings, limitations of sensing technologies or other s...
research
11/20/2016

Linear Thompson Sampling Revisited

We derive an alternative proof for the regret of Thompson sampling () in...
research
02/06/2020

Regret analysis of the Piyavskii-Shubert algorithm for global Lipschitz optimization

We consider the problem of maximizing a non-concave Lipschitz multivaria...
research
02/26/2015

A Chaining Algorithm for Online Nonparametric Regression

We consider the problem of online nonparametric regression with arbitrar...
research
04/19/2016

Streaming Label Learning for Modeling Labels on the Fly

It is challenging to handle a large volume of labels in multi-label lear...
research
06/01/2019

Sparsity Normalization: Stabilizing the Expected Outputs of Deep Networks

The learning of deep models, in which a numerous of parameters are super...

Please sign up or login with your details

Forgot password? Click here to reset