Open Access
September 2017 A Bayes Interpretation of Stacking for M-Complete and M-Open Settings
Tri Le, Bertrand Clarke
Bayesian Anal. 12(3): 807-829 (September 2017). DOI: 10.1214/16-BA1023

Abstract

In M-open problems where no true model can be conceptualized, it is common to back off from modeling and merely seek good prediction. Even in M-complete problems, taking a predictive approach can be very useful. Stacking is a model averaging procedure that gives a composite predictor by combining individual predictors from a list of models using weights that optimize a cross-validation criterion. We show that the stacking weights also asymptotically minimize a posterior expected loss. Hence we formally provide a Bayesian justification for cross-validation. Often the weights are constrained to be positive and sum to one. For greater generality, we omit the positivity constraint and relax the ‘sum to one’ constraint.

A key question is ‘What predictors should be in the average?’ We first verify that the stacking error depends only on the span of the models. Then we propose using bootstrap samples from the data to generate empirical basis elements that can be used to form models. We use this in two computed examples to give stacking predictors that are (i) data driven, (ii) optimal with respect to the number of component predictors, and (iii) optimal with respect to the weight each predictor gets.

Citation

Download Citation

Tri Le. Bertrand Clarke. "A Bayes Interpretation of Stacking for M-Complete and M-Open Settings." Bayesian Anal. 12 (3) 807 - 829, September 2017. https://doi.org/10.1214/16-BA1023

Information

Published: September 2017
First available in Project Euclid: 7 September 2016

zbMATH: 1384.62298
MathSciNet: MR3655877
Digital Object Identifier: 10.1214/16-BA1023

Keywords: Bayes action , cross-validation , optimization constrains , prediction , problem classes , stacking

Vol.12 • No. 3 • September 2017
Back to Top