SOTAVerified

Concentration inequalities for leave-one-out cross validation

2022-11-04Unverified0· sign in to hype

Benny Avelin, Lauri Viitasaari

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this article we prove that estimator stability is enough to show that leave-one-out cross validation is a sound procedure, by providing concentration bounds in a general framework. In particular, we provide concentration bounds beyond Lipschitz continuity assumptions on the loss or on the estimator. We obtain our results by relying on random variables with distribution satisfying the logarithmic Sobolev inequality, providing us a relatively rich class of distributions. We illustrate our method by considering several interesting examples, including linear regression, kernel density estimation, and stabilized/truncated estimators such as stabilized kernel regression.

Tasks

Reproductions