Conditional regression for the Nonlinear Single-Variable Model

arXiv:2411.09686v3 Announce Type: replace
Abstract: Regressing a function $F$ on $mathbb{R}^d$ without the statistical and computational curse of dimensionality requires special statistical models, for example that impose geometric assumptions on the distribution of the data (e.g., that its support is low-dimensional), or strong smoothness assumptions on $F$, or a special structure $F$. Among the latter, compositional models $F=fcirc g$ with $g$ mapping to $mathbb{R}^r$ with $rll d$ include classical single- and multi-index models, as well as neural networks. While the case where $g$ is linear is well-understood, less is known when $g$ is nonlinear, and in particular for which $g$’s the curse of dimensionality in estimating $F$, or both $f$ and $g$, may be circumvented. Here we consider a model $F(X):=f(Pi_gamma X)$ where $Pi_gamma:mathbb{R}^dto[0,textrm{len}_gamma]$ is the closest-point projection onto the parameter of a regular curve $gamma:[0, textrm{len}_gamma]tomathbb{R}^d$, and $f:[0,textrm{len}_gamma]to mathbb{R}^1$. The input data $X$ is not low-dimensional: it can be as far from $gamma$ as the condition that $Pi_gamma(X)$ is well-defined allows. The distribution $X$, the curve $gamma$ and the function $f$ are all unknown. This model is a natural nonlinear generalization of the single-index model, corresponding to $gamma$ being a line. We propose a nonparametric estimator, based on conditional regression, that under suitable assumptions, the strongest of which being that $f$ is coarsely monotone, achieves, up to log factors, the $textit{one-dimensional}$ optimal min-max rate for non-parametric regression, up to the level of noise in the observations, and be constructed in time $mathcal{O}(d^2 nlog n)$. All the constants in the learning bounds, in the minimal number of samples required for our bounds to hold, and in the computational complexity are at most low-order polynomials in $d$.

Liked Liked