Abstract
In high dimensional model selection problems, penalized least square approaches have been extensively used. The paper addresses the question of both robustness and efficiency of penalized model selection methods and proposes a data-driven weighted linear combination of convex loss functions, together with weighted L1-penalty. It is completely data adaptive and does not require prior knowledge of the error distribution. The weighted L1-penalty is used both to ensure the convexity of the penalty term and to ameliorate the bias that is caused by the L1-penalty. In the setting with dimensionality much larger than the sample size, we establish a strong oracle property of the method proposed that has both the model selection consistency and estimation efficiency for the true non-zero coefficients. As specific examples, we introduce a robust method of composite L1-L2, and an optimal composite quantile method and evaluate their performance in both simulated and real data examples.
Original language | English (US) |
---|---|
Pages (from-to) | 325-349 |
Number of pages | 25 |
Journal | Journal of the Royal Statistical Society. Series B: Statistical Methodology |
Volume | 73 |
Issue number | 3 |
DOIs | |
State | Published - Jun 2011 |
All Science Journal Classification (ASJC) codes
- Statistics and Probability
- Statistics, Probability and Uncertainty
Keywords
- Composite quasi-maximum likelihood estimation
- Lasso
- Model selection
- Non-polynomial dimensionality
- Oracle property
- Robust statistics
- Smoothly clipped absolute deviation