Sparse bridge estimation with a diverging number of parameters

Sunghoon Kwon, Yongdai Kim, Hosik Choi

Research output: Contribution to journalArticlepeer-review

12 Scopus citations

Abstract

The Bridge estimator with lνν -penalty for some ν > 0 is one of the popular choices in penalized linear regression models. It is known that, when ν = 1, the Bridge estimator produces sparse models which allow us to control the model complexity. However, when ν = 1, the Bridge estimator fails to identify the correct model since it requires certain strong sufficient conditions that are hard to hold in general, and when ν > 1, it achieves no sparsity in parameter estimation. In this paper, we propose the sparse Bridge estimator that is developed to find the correct sparse version of the Bridge estimator when ν≥ 1. Theoretically, the sparse Bridge estimator is asymptotically equivalent to the oracle Bridge estimator when the number of predictive variables diverges to infinity but less than the sample size. Here, the oracle Bridge estimator is an ideal Bridge estimator obtained by deleting all irrelevant predictive variables in advance. Hence, the sparse Bridge estimator naturally inherits the properties of the Bridge estimator without losing correct model identification asymptotically. Numerical studies show that the sparse Bridge estimator can outperform other penalized estimators with a finite sample.

Original languageEnglish
Pages (from-to)231-242
Number of pages12
JournalStatistics and its Interface
Volume6
Issue number2
DOIs
StatePublished - 2013

Keywords

  • Bridge
  • Diverging number of parameters
  • Lasso
  • Regression
  • Ridge
  • Variable selection

Fingerprint

Dive into the research topics of 'Sparse bridge estimation with a diverging number of parameters'. Together they form a unique fingerprint.

Cite this