Adaptive Explicit Kernel Minkowski Weighted K-means

12/04/2020
by   Amir Aradnia, et al.
0

The K-means algorithm is among the most commonly used data clustering methods. However, the regular K-means can only be applied in the input space and it is applicable when clusters are linearly separable. The kernel K-means, which extends K-means into the kernel space, is able to capture nonlinear structures and identify arbitrarily shaped clusters. However, kernel methods often operate on the kernel matrix of the data, which scale poorly with the size of the matrix or suffer from the high clustering cost due to the repetitive calculations of kernel values. Another issue is that algorithms access the data only through evaluations of K(x_i, x_j), which limits many processes that can be done on data through the clustering task. This paper proposes a method to combine the advantages of the linear and nonlinear approaches by using driven corresponding approximate finite-dimensional feature maps based on spectral analysis. Applying approximate finite-dimensional feature maps were only discussed in the Support Vector Machines (SVM) problems before. We suggest using this method in kernel K-means era as alleviates storing huge kernel matrix in memory, further calculating cluster centers more efficiently and access the data explicitly in feature space. These explicit feature maps enable us to access the data in the feature space explicitly and take advantage of K-means extensions in that space. We demonstrate our Explicit Kernel Minkowski Weighted K-mean (Explicit KMWK-mean) method is able to be more adopted and find best-fitting values in new space by applying additional Minkowski exponent and feature weights parameter. Moreover, it can reduce the impact of concentration on nearest neighbour search by suggesting investigate among other norms instead of Euclidean norm, includes Minkowski norms and fractional norms (as an extension of the Minkowski norms with p<1).

READ FULL TEXT
research
12/14/2018

Class Mean Vector Component and Discriminant Analysis for Kernel Subspace Learning

The kernel matrix used in kernel methods encodes all the information req...
research
03/12/2015

Compact Nonlinear Maps and Circulant Extensions

Kernel approximation via nonlinear random feature maps is widely used in...
research
11/29/2012

Classification Recouvrante Basée sur les Méthodes à Noyau

Overlapping clustering problem is an important learning issue in which c...
research
10/09/2018

Learning Bounds for Greedy Approximation with Explicit Feature Maps from Multiple Kernels

Nonlinear kernels can be approximated using finite-dimensional feature m...
research
12/27/2019

Nonlinear Markov Clustering by Minimum Curvilinear Sparse Similarity

The development of algorithms for unsupervised pattern recognition by no...
research
06/18/2020

Quantum algorithm for training nonlinear SVMs in almost linear time

We propose a quantum algorithm for training nonlinear support vector mac...
research
07/02/2018

Multi-distance Support Matrix Machines

Real-world data such as digital images, MRI scans and electroencephalogr...

Please sign up or login with your details

Forgot password? Click here to reset