Linear Time Kernel Matrix Approximation via Hyperspherical Harmonics

02/08/2022
by   John Paul Ryan, et al.
0

We propose a new technique for constructing low-rank approximations of matrices that arise in kernel methods for machine learning. Our approach pairs a novel automatically constructed analytic expansion of the underlying kernel function with a data-dependent compression step to further optimize the approximation. This procedure works in linear time and is applicable to any isotropic kernel. Moreover, our method accepts the desired error tolerance as input, in contrast to prevalent methods which accept the rank as input. Experimental results show our approach compares favorably to the commonly used Nystrom method with respect to both accuracy for a given rank and computational time for a given accuracy across a variety of kernels, dimensions, and datasets. Notably, in many of these problem settings our approach produces near-optimal low-rank approximations. We provide an efficient open-source implementation of our new technique to complement our theoretical developments and experimental results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/27/2021

Efficient randomized tensor-based algorithms for function approximation and low-rank kernel interactions

In this paper, we introduce a method for multivariate function approxima...
research
09/05/2018

IKA: Independent Kernel Approximator

This paper describes a new method for low rank kernel approximation call...
research
02/11/2020

Generalization Guarantees for Sparse Kernel Approximation with Entropic Optimal Features

Despite their success, kernel methods suffer from a massive computationa...
research
06/27/2023

FLuRKA: Fast fused Low-Rank Kernel Attention

Many efficient approximate self-attention techniques have become prevale...
research
05/03/2015

Structured Block Basis Factorization for Scalable Kernel Matrix Evaluation

Kernel matrices are popular in machine learning and scientific computing...
research
03/14/2019

Low-rank Kernel Learning for Graph-based Clustering

Constructing the adjacency graph is fundamental to graph-based clusterin...
research
06/12/2020

Linear Time Sinkhorn Divergences using Positive Features

Although Sinkhorn divergences are now routinely used in data sciences to...

Please sign up or login with your details

Forgot password? Click here to reset