Translator Disclaimer
December 2014 A Stochastic Variational Framework for Fitting and Diagnosing Generalized Linear Mixed Models
Linda S. L. Tan, David J. Nott
Bayesian Anal. 9(4): 963-1004 (December 2014). DOI: 10.1214/14-BA885


In stochastic variational inference, the variational Bayes objective function is optimized using stochastic gradient approximation, where gradients computed on small random subsets of data are used to approximate the true gradient over the whole data set. This enables complex models to be fit to large data sets as data can be processed in mini-batches. In this article, we extend stochastic variational inference for conjugate-exponential models to nonconjugate models and present a stochastic nonconjugate variational message passing algorithm for fitting generalized linear mixed models that is scalable to large data sets. In addition, we show that diagnostics for prior-likelihood conflict, which are useful for Bayesian model criticism, can be obtained from nonconjugate variational message passing automatically, as an alternative to simulation-based Markov chain Monte Carlo methods. Finally, we demonstrate that for moderate-sized data sets, convergence can be accelerated by using the stochastic version of nonconjugate variational message passing in the initial stage of optimization before switching to the standard version.


Download Citation

Linda S. L. Tan. David J. Nott. "A Stochastic Variational Framework for Fitting and Diagnosing Generalized Linear Mixed Models." Bayesian Anal. 9 (4) 963 - 1004, December 2014.


Published: December 2014
First available in Project Euclid: 21 November 2014

zbMATH: 1327.62167
MathSciNet: MR3293964
Digital Object Identifier: 10.1214/14-BA885

Rights: Copyright © 2014 International Society for Bayesian Analysis


Vol.9 • No. 4 • December 2014
Back to Top