SOTAVerified

Tight Kernel Query Complexity of Kernel Ridge Regression and Kernel k-means Clustering

2019-05-15Unverified0· sign in to hype

Manuel Fernandez, David P. Woodruff, Taisuke Yasuda

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We present tight lower bounds on the number of kernel evaluations required to approximately solve kernel ridge regression (KRR) and kernel k-means clustering (KKMC) on n input points. For KRR, our bound for relative error approximation to the minimizer of the objective function is (nd_eff^/) where d_eff^ is the effective statistical dimension, which is tight up to a (d_eff^/) factor. For KKMC, our bound for finding a k-clustering achieving a relative error approximation of the objective function is (nk/), which is tight up to a (k/) factor. Our KRR result resolves a variant of an open question of El Alaoui and Mahoney, asking whether the effective statistical dimension is a lower bound on the sampling complexity or not. Furthermore, for the important practical case when the input is a mixture of Gaussians, we provide a KKMC algorithm which bypasses the above lower bound.

Tasks

Reproductions