Adaptive Sampling for Coarse Ranking

02/20/2018
by   Sumeet Katariya, et al.
0

We consider the problem of active coarse ranking, where the goal is to sort items according to their means into clusters of pre-specified sizes, by adaptively sampling from their reward distributions. This setting is useful in many social science applications involving human raters and the approximate rank of every item is desired. Approximate or coarse ranking can significantly reduce the number of ratings required in comparison to the number needed to find an exact ranking. We propose a computationally efficient PAC algorithm LUCBRank for coarse ranking, and derive an upper bound on its sample complexity. We also derive a nearly matching distribution-dependent lower bound. Experiments on synthetic as well as real-world data show that LUCBRank performs better than state-of-the-art baseline methods, even when these methods have the advantage of knowing the underlying parametric model.

READ FULL TEXT
research
06/08/2018

PAC Ranking from Pairwise and Listwise Queries: Lower Bounds and Upper Bounds

This paper explores the adaptively (active) PAC (probably approximately ...
research
10/23/2018

Active Ranking with Subset-wise Preferences

We consider the problem of probably approximately correct (PAC) ranking ...
research
08/11/2018

Ranking with Features: Algorithm and A Graph Theoretic Analysis

We consider the problem of ranking a set of items from pairwise comparis...
research
09/07/2019

On Sample Complexity Upper and Lower Bounds for Exact Ranking from Noisy Comparisons

This paper studies the problem of finding the exact ranking from noisy c...
research
11/02/2021

Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification

We study the problem of the identification of m arms with largest means ...
research
11/15/2018

Secretary Ranking with Minimal Inversions

We study a twist on the classic secretary problem, which we term the sec...
research
04/15/2018

Higher-order Coreference Resolution with Coarse-to-fine Inference

We introduce a fully differentiable approximation to higher-order infere...

Please sign up or login with your details

Forgot password? Click here to reset