Publication: L1-Regularized Least Squares for Support Recovery of High Dimensional Single Index Models with Gaussian Designs
Open/View Files
Date
2016
Authors
Published Version
Published Version
Journal Title
Journal ISSN
Volume Title
Publisher
The Harvard community has made this article openly available. Please share how this access benefits you.
Citation
Neykov, Matey, Jun S. Liu, and Tianxi Cai. 2016. "L1-regularized least squares for support recovery of high dimensional single index models with Gaussian designs." Journal of Machine Learning Research 17, no. 87: 1-37.
Research Data
Abstract
It is known that for a certain class of single index models (SIMs) z[]c0, support recovery is impossible when X ~ [N](0, []p×p) and a model complexity adjusted sample size is below a critical threshold. Recently, optimal algorithms based on Sliced Inverse Regression (SIR) were suggested. These algorithms work provably under the assumption that the design X comes from an i.i.d. Gaussian distribution. In the present paper we analyze algorithms based on covariance screening and least squares with L1 penalization (i.e. LASSO) and demonstrate that they can also enjoy optimal (up to a scalar) rescaled sample size in terms of support recovery, albeit under slightly different assumptions on f and [e] compared to the SIR based algorithms. Furthermore, we show more generally, that LASSO succeeds in recovering the signed support of β0 if X ~ [N] (0, [Sigma]), and the covariance [Sigma] satisfies the irrepresentable condition. Our work extends existing results on the support recovery of LASSO for the linear model, to a more general class of SIMs.
Description
Other Available Sources
Keywords
Single index models, Sparsity, Support recovery, High-dimensional statistics, LASSO
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service