On the identifiability of mixtures of ranking models

by   Xiaomin Zhang, et al.

Mixtures of ranking models are standard tools for ranking problems. However, even the fundamental question of parameter identifiability is not fully understood: the identifiability of a mixture model with two Bradley-Terry-Luce (BTL) components has remained open. In this work, we show that popular mixtures of ranking models with two components (Plackett-Luce, multinomial logistic model with slates of size 3, or BTL) are generically identifiable, i.e., the ground-truth parameters can be identified except when they are from a pathological subset of measure zero. We provide a framework for verifying the number of solutions in a general family of polynomial systems using algebraic geometry, and apply it to these mixtures of ranking models. The framework can be applied more broadly to other learning models and may be of independent interest.


page 20

page 23

page 24

page 31

page 34

page 36

page 37

page 42


Mixtures of Experts Models

Mixtures of experts models provide a framework in which covariates may b...

Training Mixture Models at Scale via Coresets

How can we train a statistical mixture model on a massive data set? In t...

Efficiently Learning Mixtures of Mallows Models

Mixtures of Mallows models are a popular generative model for ranking da...

Learning Mixtures of Plackett-Luce Models from Structured Partial Orders

Mixtures of ranking models have been widely used for heterogeneous prefe...

Finite and infinite Mallows ranking models, maximum likelihood estimator, and regeneration

In this paper we are concerned with various Mallows ranking models. Firs...

Concentric mixtures of Mallows models for top-k rankings: sampling and identifiability

In this paper, we consider mixtures of two Mallows models for top-k rank...

Mixtures and products in two graphical models

We compare two statistical models of three binary random variables. One ...

Please sign up or login with your details

Forgot password? Click here to reset