Open Access
Translator Disclaimer
March 2015 Power-Expected-Posterior Priors for Variable Selection in Gaussian Linear Models
Dimitris Fouskakis, Ioannis Ntzoufras, David Draper
Bayesian Anal. 10(1): 75-107 (March 2015). DOI: 10.1214/14-BA887

Abstract

In the context of the expected-posterior prior (EPP) approach to Bayesian variable selection in linear models, we combine ideas from power-prior and unit-information-prior methodologies to simultaneously (a) produce a minimally-informative prior and (b) diminish the effect of training samples. The result is that in practice our power-expected-posterior (PEP) methodology is sufficiently insensitive to the size n* of the training sample, due to PEP’s unit-information construction, that one may take n* equal to the full-data sample size n and dispense with training samples altogether. This promotes stability of the resulting Bayes factors, removes the arbitrariness arising from individual training-sample selections, and greatly increases computational speed, allowing many more models to be compared within a fixed CPU budget. We find that, under an independence Jeffreys (reference) baseline prior, the asymptotics of PEP Bayes factors are equivalent to those of Schwartz’s Bayesian Information Criterion (BIC), ensuring consistency of the PEP approach to model selection. Our PEP prior, due to its unit-information structure, leads to a variable-selection procedure that — in our empirical studies — (1) is systematically more parsimonious than the basic EPP with minimal training sample, while sacrificing no desirable performance characteristics to achieve this parsimony; (2) is robust to the size of the training sample, thus enjoying the advantages described above arising from the avoidance of training samples altogether; and (3) identifies maximum-a-posteriori models that achieve better out-of-sample predictive performance than that provided by standard EPPs, the g-prior, the hyper-g prior, non-local priors, the Least Absolute Shrinkage and Selection Operator (LASSO) and Smoothly-Clipped Absolute Deviation (SCAD) methods.

Citation

Download Citation

Dimitris Fouskakis. Ioannis Ntzoufras. David Draper. "Power-Expected-Posterior Priors for Variable Selection in Gaussian Linear Models." Bayesian Anal. 10 (1) 75 - 107, March 2015. https://doi.org/10.1214/14-BA887

Information

Published: March 2015
First available in Project Euclid: 28 January 2015

zbMATH: 1335.62045
MathSciNet: MR3420898
Digital Object Identifier: 10.1214/14-BA887

Keywords: Bayes factors , Bayesian variable selection , consistency , expected-posterior priors , Gaussian linear models , g-prior , Hyper-g prior , Lasso , Non-local priors , Power-prior , Prior compatibility , SCAD , training samples , unit-information prior

Rights: Copyright © 2015 International Society for Bayesian Analysis

JOURNAL ARTICLE
33 PAGES


SHARE
Vol.10 • No. 1 • March 2015
Back to Top