SOTAVerified

From PAC-Bayes Bounds to KL Regularization

2009-12-01NeurIPS 2009Unverified0· sign in to hype

Pascal Germain, Alexandre Lacasse, Mario Marchand, Sara Shanian, François Laviolette

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We show that convex KL-regularized objective functions are obtained from a PAC-Bayes risk bound when using convex loss functions for the stochastic Gibbs classifier that upper-bound the standard zero-one loss used for the weighted majority vote. By restricting ourselves to a class of posteriors, that we call quasi uniform, we propose a simple coordinate descent learning algorithm to minimize the proposed KL-regularized cost function. We show that standard ell_p-regularized objective functions currently used, such as ridge regression and ell_p-regularized boosting, are obtained from a relaxation of the KL divergence between the quasi uniform posterior and the uniform prior. We present numerical experiments where the proposed learning algorithm generally outperforms ridge regression and AdaBoost.

Tasks

Reproductions