首页    期刊浏览 2024年09月19日 星期四
登录注册

文章基本信息

  • 标题:On the Asymptotic Properties of SLOPE
  • 本地全文:下载
  • 作者:Michał Kos ; Małgorzata Bogdan
  • 期刊名称:Sankhya. Series A, mathematical statistics and probability
  • 印刷版ISSN:0976-836X
  • 电子版ISSN:0976-8378
  • 出版年度:2020
  • 卷号:82
  • 期号:2
  • 页码:499-532
  • DOI:10.1007/s13171-020-00212-5
  • 出版社:Indian Statistical Institute
  • 摘要:Abstract Sorted L-One Penalized Estimator (SLOPE) is a relatively new convex optimization procedure for selecting predictors in high dimensional regression analyses. SLOPE extends LASSO by replacing the L 1 penalty norm with a Sorted L 1 norm, based on the non-increasing sequence of tuning parameters. This allows SLOPE to adapt to unknown sparsity and achieve an asymptotic minimax convergency rate under a wide range of high dimensional generalized linear models. Additionally, in the case when the design matrix is orthogonal, SLOPE with the sequence of tuning parameters λ B H corresponding to the sequence of decaying thresholds for the Benjamini-Hochberg multiple testing correction provably controls the False Discovery Rate (FDR) in the multiple regression model. In this article we provide new asymptotic results on the properties of SLOPE when the elements of the design matrix are iid random variables from the Gaussian distribution. Specifically, we provide conditions under which the asymptotic FDR of SLOPE based on the sequence λ B H converges to zero and the power converges to 1. We illustrate our theoretical asymptotic results with an extensive simulation study. We also provide precise formulas describing FDR of SLOPE under different loss functions, which sets the stage for future investigation on the model selection properties of SLOPE and its extensions.
  • 关键词:Multiple testing;Model selection;High dimensional regression;Convex optimization.
国家哲学社会科学文献中心版权所有