Oversampling Divide-and-conquer for Response-skewed Kernel Ridge Regression

07/13/2021
by   Jingyi Zhang, et al.
0

The divide-and-conquer method has been widely used for estimating large-scale kernel ridge regression estimates. Unfortunately, when the response variable is highly skewed, the divide-and-conquer kernel ridge regression (dacKRR) may overlook the underrepresented region and result in unacceptable results. We develop a novel response-adaptive partition strategy to overcome the limitation. In particular, we propose to allocate the replicates of some carefully identified informative observations to multiple nodes (local processors). The idea is analogous to the popular oversampling technique. Although such a technique has been widely used for addressing discrete label skewness, extending it to the dacKRR setting is nontrivial. We provide both theoretical and practical guidance on how to effectively over-sample the observations under the dacKRR setting. Furthermore, we show the proposed estimate has a smaller asymptotic mean squared error (AMSE) than that of the classical dacKRR estimate under mild conditions. Our theoretical findings are supported by both simulated and real-data analyses.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2021

Uncertainty quantification for distributed regression

The ever-growing size of the datasets renders well-studied learning tech...
research
08/05/2016

Kernel Ridge Regression via Partitioning

In this paper, we investigate a divide and conquer approach to Kernel Ri...
research
06/30/2020

Optimal Rates of Distributed Regression with Imperfect Kernels

Distributed machine learning systems have been receiving increasing atte...
research
09/08/2023

Adaptive Distributed Kernel Ridge Regression: A Feasible Distributed Learning Scheme for Data Silos

Data silos, mainly caused by privacy and interoperability, significantly...
research
05/04/2020

Reduced Rank Multivariate Kernel Ridge Regression

In the multivariate regression, also referred to as multi-task learning ...
research
04/18/2019

Ridge regularization for Mean Squared Error Reduction in Regression with Weak Instruments

In this paper, I show that classic two-stage least squares (2SLS) estima...
research
04/24/2023

Robust, randomized preconditioning for kernel ridge regression

This paper introduces two randomized preconditioning techniques for robu...

Please sign up or login with your details

Forgot password? Click here to reset