SOTAVerified

Second-Order Bounds for [0,1]-Valued Regression via Betting Loss

2025-07-16Unverified0· sign in to hype

Yinan Li, Kwang-Sung Jun

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We consider the [0,1]-valued regression problem in the i.i.d. setting. In a related problem called cost-sensitive classification, foster21efficient have shown that the log loss minimizer achieves an improved generalization bound compared to that of the squared loss minimizer in the sense that the bound scales with the cost of the best classifier, which can be arbitrarily small depending on the problem at hand. Such a result is often called a first-order bound. For [0,1]-valued regression, we first show that the log loss minimizer leads to a similar first-order bound. We then ask if there exists a loss function that achieves a variance-dependent bound (also known as a second order bound), which is a strict improvement upon first-order bounds. We answer this question in the affirmative by proposing a novel loss function called the betting loss. Our result is ``variance-adaptive'' in the sense that the bound is attained without any knowledge about the variance, which is in contrast to modeling label (or reward) variance or the label distribution itself explicitly as part of the function class such as distributional reinforcement learning.

Tasks

Reproductions