Relative Deviation Learning Bounds and Generalization with Unbounded Loss Functions
2013-10-22Unverified0· sign in to hype
Corinna Cortes, Spencer Greenberg, Mehryar Mohri
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We present an extensive analysis of relative deviation bounds, including detailed proofs of two-sided inequalities and their implications. We also give detailed proofs of two-sided generalization bounds that hold in the general case of unbounded loss functions, under the assumption that a moment of the loss is bounded. These bounds are useful in the analysis of importance weighting and other learning tasks such as unbounded regression.