SOTAVerified

Nonparametric Greedy Algorithms for the Sparse Learning Problem

2009-12-01NeurIPS 2009Unverified0· sign in to hype

Han Liu, Xi Chen

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper studies the forward greedy strategy in sparse nonparametric regression. For additive models, we propose an algorithm called additive forward regression; for general multivariate regression, we propose an algorithm called generalized forward regression. Both of them simultaneously conduct estimation and variable selection in nonparametric settings for the high dimensional sparse learning problem. Our main emphasis is empirical: on both simulated and real data, these two simple greedy methods can clearly outperform several state-of-the-art competitors, including the LASSO, a nonparametric version of the LASSO called the sparse additive model (SpAM) and a recently proposed adaptive parametric forward-backward algorithm called the Foba. Some theoretical justifications are also provided.

Tasks

Reproductions