Conditional regression for the Nonlinear Single-Variable Model
Yantao Wu, Mauro Maggioni
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Several statistical models for regression of a function F on R^d without the statistical and computational curse of dimensionality exist, for example by imposing and exploiting geometric assumptions on the distribution of the data (e.g. that its support is low-dimensional), or strong smoothness assumptions on F, or a special structure F. Among the latter, compositional models assume F=f g with g mapping to R^r with r d, have been studied, and include classical single- and multi-index models and recent works on neural networks. While the case where g is linear is rather well-understood, much less is known when g is nonlinear, and in particular for which g's the curse of dimensionality in estimating F, or both f and g, may be circumvented. In this paper, we consider a model F(X):=f(_ X) where _:R^d[0,len_] is the closest-point projection onto the parameter of a regular curve : [0,len_]R^d and f:[0,len_]R^1. The input data X is not low-dimensional, far from , conditioned on _(X) being well-defined. The distribution of the data, and f are unknown. This model is a natural nonlinear generalization of the single-index model, which corresponds to being a line. We propose a nonparametric estimator, based on conditional regression, and show that under suitable assumptions, the strongest of which being that f is coarsely monotone, it can achieve the one-dimensional optimal min-max rate for non-parametric regression, up to the level of noise in the observations, and be constructed in time O(d^2n n). All the constants in the learning bounds, in the minimal number of samples required for our bounds to hold, and in the computational complexity are at most low-order polynomials in d.