Abstract
Regression models with crossed random effect errors can be very expensive to compute. The cost of both generalized least squares and Gibbs sampling can easily grow as (or worse) for N observations. Papaspiliopoulos, Roberts and Zanella (Biometrika 107 (2020) 25–40) present a collapsed Gibbs sampler that costs , but under an extremely stringent sampling model. We propose a backfitting algorithm to compute a generalized least squares estimate and prove that it costs . A critical part of the proof is in ensuring that the number of iterations required is , which follows from keeping a certain matrix norm below for some . Our conditions are greatly relaxed compared to those for the collapsed Gibbs sampler, though still strict. Empirically, the backfitting algorithm has a norm below under conditions that are less strict than those in our assumptions. We illustrate the new algorithm on a ratings data set from Stitch Fix.
Funding Statement
This work was supported by the U.S. National Science Foundation under Grant IIS-1837931.
Acknowledgments
We are grateful to Brad Klingenberg and Stitch Fix for sharing some test data with us and James Johndrow for some helpful discussions. We thank the reviewers for remarks that have helped us improve the paper.
Citation
Swarnadip Ghosh. Trevor Hastie. Art B. Owen. "Backfitting for large scale crossed random effects regressions." Ann. Statist. 50 (1) 560 - 583, February 2022. https://doi.org/10.1214/21-AOS2121
Information