Abstract
Several sparseness penalties have been suggested for delivery of good predictive performance in automatic variable selection within the framework of regularization. All assume that the true model is sparse. We propose a penalty, a convex combination of the L1- and L∞-norms, that adapts to a variety of situations including sparseness and nonsparseness, grouping and nongrouping. The proposed penalty performs grouping and adaptive regularization. In addition, we introduce a novel homotopy algorithm utilizing subgradients for developing regularization solution surfaces involving multiple regularizers. This permits efficient computation and adaptive tuning. Numerical experiments are conducted using simulation. In simulated and real examples, the proposed penalty compares well against popular alternatives.
Original language | English (US) |
---|---|
Pages (from-to) | 513-527 |
Number of pages | 15 |
Journal | Biometrika |
Volume | 96 |
Issue number | 3 |
DOIs | |
State | Published - Sep 2009 |
Bibliographical note
Funding Information:This research was supported in part by grants from the U.S. National Science Foundation and National Institutes of Health.
Keywords
- Homotopy
- L1-norm
- Lasso
- L∞-norm
- Subgradient
- Support vector machine
- Variable grouping and selection