Abstract
This paper concerns the robust regression model when the number of predictors and the number of observations grow in a similar rate. Theory for M-estimators in this regime has been recently developed by several authors (El Karoui et al., 2013; Bean et al., 2013; Donoho and Montanari, 2013). Motivated by the inability of M-estimators to successfully estimate the Euclidean norm of the coefficient vector, we consider a Bayesian framework for this model. We suggest a two-component mixture of normals prior for the coefficients and develop a Gibbs sampler procedure for sampling from relevant posterior distributions, while utilizing a scale mixture of normal representation for the error distribution. Unlike M-estimators, the proposed Bayes estimator is consistent in the Euclidean norm sense. Simulation results demonstrate the superiority of the Bayes estimator over traditional estimation methods.
| Original language | English |
|---|---|
| Pages (from-to) | 3045-3062 |
| Number of pages | 18 |
| Journal | Electronic Journal of Statistics |
| Volume | 10 |
| Issue number | 2 |
| DOIs | |
| State | Published - 2016 |
Bibliographical note
Publisher Copyright:© 2016, Institute of Mathematical Statistics. All rights reserved.
Keywords
- Bayesian estimation
- High dimensional regression
- MCMC
- Robust regression
Fingerprint
Dive into the research topics of 'On bayesian robust regression with diverging number of predictors'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver