On Data-Dependent Random Features for Improved Generalization in Supervised Learning

12/19/2017
by   Shahin Shahrampour, et al.
0

The randomized-feature approach has been successfully employed in large-scale kernel approximation and supervised learning. The distribution from which the random features are drawn impacts the number of features required to efficiently perform a learning task. Recently, it has been shown that employing data-dependent randomization improves the performance in terms of the required number of random features. In this paper, we are concerned with the randomized-feature approach in supervised learning for good generalizability. We propose the Energy-based Exploration of Random Features (EERF) algorithm based on a data-dependent score function that explores the set of possible features and exploits the promising regions. We prove that the proposed score function with high probability recovers the spectrum of the best fit within the model class. Our empirical results on several benchmark datasets further verify that our method requires smaller number of random features to achieve a certain generalization error compared to the state-of-the-art while introducing negligible pre-processing overhead. EERF can be implemented in a few lines of code and requires no additional tuning parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/20/2019

On Sampling Random Features From Empirical Leverage Scores: Implementation and Theoretical Guarantees

Random features provide a practical framework for large-scale kernel app...
research
11/27/2017

Data Dependent Kernel Approximation using Pseudo Random Fourier Features

Kernel methods are powerful and flexible approach to solve many problems...
research
10/11/2019

A General Scoring Rule for Randomized Kernel Approximation with Application to Canonical Correlation Analysis

Random features has been widely used for kernel approximation in large-s...
research
05/23/2017

Data-driven Random Fourier Features using Stein Effect

Large-scale kernel approximation is an important problem in machine lear...
research
10/09/2018

Learning Bounds for Greedy Approximation with Explicit Feature Maps from Multiple Kernels

Nonlinear kernels can be approximated using finite-dimensional feature m...
research
02/14/2016

Generalization Properties of Learning with Random Features

We study the generalization properties of ridge regression with random f...
research
01/30/2019

Ensemble-based kernel learning for a class of data assimilation problems with imperfect forward simulators

Simulator imperfection, often known as model error, is ubiquitous in pra...

Please sign up or login with your details

Forgot password? Click here to reset