Distribution-based Label Space Transformation for Multi-label Learning

05/15/2018
by   Zongting Lyu, et al.
0

Multi-label learning problems have manifested themselves in various machine learning applications. The key to successful multi-label learning algorithms lies in the exploration of inter-label correlations, which usually incur great computational cost. Another notable factor in multi-label learning is that the label vectors are usually extremely sparse, especially when the candidate label vocabulary is very large and only a few instances are assigned to each category. Recently, a label space transformation (LST) framework has been proposed targeting these challenges. However, current methods based on LST usually suffer from information loss in the label space dimension reduction process and fail to address the sparsity problem effectively. In this paper, we propose a distribution-based label space transformation (DLST) model. By defining the distribution based on the similarity of label vectors, a more comprehensive label structure can be captured. Then, by minimizing KL-divergence of two distributions, the information of the original label space can be approximately preserved in the latent space. Consequently, multi-label classifier trained using the dense latent codes yields better performance. The leverage of distribution enables DLST to fill out additional information about the label correlations. This endows DLST the capability to handle label set sparsity and training data sparsity in multi-label learning problems. With the optimal latent code, a kernel logistic regression function is learned for the mapping from feature space to the latent space. Then ML-KNN is employed to recover the original label vector from the transformed latent code. Extensive experiments on several benchmark datasets demonstrate that DLST not only achieves high classification performance but also is computationally more efficient.

READ FULL TEXT

page 1

page 3

page 6

page 7

page 8

page 9

page 10

research
04/07/2020

Generalized Label Enhancement with Sample Correlations

Recently, label distribution learning (LDL) has drawn much attention in ...
research
11/15/2019

Multi-Label Learning with Deep Forest

In multi-label learning, each instance is associated with multiple label...
research
10/25/2022

TabMixer: Excavating Label Distribution Learning with Small-scale Features

Label distribution learning (LDL) differs from multi-label learning whic...
research
03/30/2016

Cost-Sensitive Label Embedding for Multi-Label Classification

Label embedding (LE) is an important family of multi-label classificatio...
research
05/09/2023

Minimal Learning Machine for Multi-Label Learning

Distance-based supervised method, the minimal learning machine, construc...
research
07/21/2021

Integration of Autoencoder and Functional Link Artificial Neural Network for Multi-label Classification

Multi-label (ML) classification is an actively researched topic currentl...
research
02/11/2021

EvoSplit: An evolutionary approach to split a multi-label data set into disjoint subsets

This paper presents a new evolutionary approach, EvoSplit, for the distr...

Please sign up or login with your details

Forgot password? Click here to reset