On the Asymptotic Properties of SLOPE

Sankhya A - Tập 82 Số 2 - Trang 499-532 - 2020
Michał Kos1, Małgorzata Bogdan1,2
1Department of Mathematics, University of Wroclaw, Wroclaw, Poland
2Department of Statistics, Lund University, Lund, Sweden

Tóm tắt

AbstractSorted L-One Penalized Estimator (SLOPE) is a relatively new convex optimization procedure for selecting predictors in high dimensional regression analyses. SLOPE extends LASSO by replacing the L1 penalty norm with a Sorted L1 norm, based on the non-increasing sequence of tuning parameters. This allows SLOPE to adapt to unknown sparsity and achieve an asymptotic minimax convergency rate under a wide range of high dimensional generalized linear models. Additionally, in the case when the design matrix is orthogonal, SLOPE with the sequence of tuning parameters λBH corresponding to the sequence of decaying thresholds for the Benjamini-Hochberg multiple testing correction provably controls the False Discovery Rate (FDR) in the multiple regression model. In this article we provide new asymptotic results on the properties of SLOPE when the elements of the design matrix are iid random variables from the Gaussian distribution. Specifically, we provide conditions under which the asymptotic FDR of SLOPE based on the sequence λBH converges to zero and the power converges to 1. We illustrate our theoretical asymptotic results with an extensive simulation study. We also provide precise formulas describing FDR of SLOPE under different loss functions, which sets the stage for future investigation on the model selection properties of SLOPE and its extensions.

Từ khóa


Tài liệu tham khảo

Abramovich, F. and Grinshtein, V. (2017). High-dimensional classification by sparse logistic regression. IEEE Transactions on Information Theory, PP, 06. https://doi.org/10.1109/TIT.2018.2884963.

Abramovich, F., Benjamini, Y., Donoho, D.L. and Johnstone, I.M. (2006). Adapting to unknown sparsity by controlling the false discovery rate. Annals of Statistics 34, 2, 584–653.

Bellec, P.C., Lecué, G. and Tsybakov, A.B. (2018). Slope meets lasso: improved oracle bounds and optimality. Annals of Statistics 46, 6B, 3603–3642.

Benjamini, Y. and Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society, Series B 57, 1, 289–300.

Bogdan, M., Chakrabarti, A., Frommlet, F. and Ghosh, J.K. (2011). Asymptotic Bayes optimality under sparsity of some multiple testing procedures. Annals of Statistics 39, 1551–1579.

Bogdan, M., van den Berg, E., Su, W. and Candès, E.J. (2013). Statistical estimation and testing via the ordered ℓ1 norm. Technical Report 2013-07, Department of Statistics Stanford University.

Bogdan, M., van den Berg, E., Sabatti, C., Su, W. and Candès, E.J. (2015). Slope – adaptive variable selection via convex optimization. Annals of Applied Statistics 9, 3, 1103–1140.

Candès, E.J., Wakin, M.B. and Boyd, S.P. (2008). . J. Fourier Anal. Appl. 14, 877–905.

Chen, S. and Donoho. D. (1994). Basis pursuit, 1. IEEE, p. 41–44.

Frommlet, F. and Bogdan, M. (2013). Some optimality properties of fDR controlling rules under sparsity. Electronic Journal of Statistics 7, 1328–1368.

Jiang, W., Bogdan, M., Josse, J., Miasojedow, B. and An TB Group Rockova, V. (2019). Adaptive bayesian slope-high-dimensional model selection with missing values. arXiv:1909.06631.

Neuvial, P. and Roquain, E. (2012). On false discovery rate thresholding for classification under sparsity. Annals of Statistics 40, 2572–2600.

Su, W. and Candès, E. (2016). Slope is adaptive to unknown sparsity and asymptotically minimax. Annals of Statistics 44, 3, 1038–1068,06. doi: https://doi.org/10.1214/15-AOS1397.

Su, W., Bogdan, M. and Candès, E.J. (2017). False discoveries occur early on the lasso path. Annals of Statistics 45, 5, 2133–2150.

Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), pp. 267–288.

Zou, H. (2006). The adaptive lasso and its oracle properties. Journal of the American Statistical Association 101, 476, 1418–1429. doi: https://doi.org/10.1198/016214506000000735.