Online Learning to Rank with Features

10/05/2018
by   Tor Lattimore, et al.
8

We introduce a new model for online ranking in which the click probability factors into an examination and attractiveness function and the attractiveness function is a linear function of a feature vector and an unknown parameter. Only relatively mild assumptions are made on the examination function. A novel algorithm for this setup is analysed, showing that the dependence on the number of items is replaced by a dependence on the dimension, allowing the new algorithm to handle a large number of items. When reduced to the orthogonal case, the regret of the algorithm improves on the state-of-the-art.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/29/2020

Logarithmic Regret for Adversarial Online Control

We introduce a new algorithm for online linear-quadratic control in a kn...
research
05/02/2023

Exploration of Unranked Items in Safe Online Learning to Re-Rank

Bandit algorithms for online learning to rank (OLTR) problems often aim ...
research
01/23/2019

Thompson Sampling for a Fatigue-aware Online Recommendation System

In this paper we consider an online recommendation setting, where a plat...
research
02/22/2022

No-Regret Learning in Partially-Informed Auctions

Auctions with partially-revealed information about items are broadly emp...
research
02/06/2022

Pushing the Efficiency-Regret Pareto Frontier for Online Learning of Portfolios and Quantum States

We revisit the classical online portfolio selection problem. It is widel...
research
04/20/2021

Joint Online Learning and Decision-making via Dual Mirror Descent

We consider an online revenue maximization problem over a finite time ho...
research
02/12/2018

PRIL: Perceptron Ranking Using Interval Labeled Data

In this paper, we propose an online learning algorithm PRIL for learning...

Please sign up or login with your details

Forgot password? Click here to reset