On bayesian robust regression with diverging number of predictors

Daniel Nevo, Ya’Acov Ritov

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

This paper concerns the robust regression model when the number of predictors and the number of observations grow in a similar rate. Theory for M-estimators in this regime has been recently developed by several authors (El Karoui et al., 2013; Bean et al., 2013; Donoho and Montanari, 2013). Motivated by the inability of M-estimators to successfully estimate the Euclidean norm of the coefficient vector, we consider a Bayesian framework for this model. We suggest a two-component mixture of normals prior for the coefficients and develop a Gibbs sampler procedure for sampling from relevant posterior distributions, while utilizing a scale mixture of normal representation for the error distribution. Unlike M-estimators, the proposed Bayes estimator is consistent in the Euclidean norm sense. Simulation results demonstrate the superiority of the Bayes estimator over traditional estimation methods.

Original languageEnglish
Pages (from-to)3045-3062
Number of pages18
JournalElectronic Journal of Statistics
Volume10
Issue number2
DOIs
StatePublished - 2016

Bibliographical note

Publisher Copyright:
© 2016, Institute of Mathematical Statistics. All rights reserved.

Keywords

  • Bayesian estimation
  • High dimensional regression
  • MCMC
  • Robust regression

Fingerprint

Dive into the research topics of 'On bayesian robust regression with diverging number of predictors'. Together they form a unique fingerprint.

Cite this