TY - JOUR
T1 - Approximate Cross-Validated Mean Estimates for Bayesian Hierarchical Regression Models
AU - Zhang, Amy
AU - Daniels, Michael J.
AU - Li, Changcheng
AU - Bao, Le
N1 - Publisher Copyright:
© 2024 The Author(s). Published with license by Taylor & Francis Group, LLC.
PY - 2024
Y1 - 2024
N2 - We introduce a novel procedure for obtaining cross-validated predictive estimates for Bayesian hierarchical regression models (BHRMs). BHRMs are popular for modeling complex dependence structures (e.g., Gaussian processes and Gaussian Markov random fields) but can be computationally expensive to run. Cross-validation (CV) is, therefore, not a common practice to evaluate the predictive performance of BHRMs. Our method circumvents the need to rerun computationally costly estimation methods for each cross-validation fold and makes CV more feasible for large BHRMs. We shift the CV problem from probability-based sampling to a familiar and straightforward optimization problem by conditioning on the variance-covariance parameters. Our approximation applies to leave-one-out CV and leave-one-cluster-out CV, the latter of which is more appropriate for models with complex dependencies. In many cases, this produces estimates equivalent to full CV. We provide theoretical results, demonstrate the efficacy of our method on publicly available data and in simulations, and compare the model performance with several competing methods for CV approximation. Code and other supplementary materials available online.
AB - We introduce a novel procedure for obtaining cross-validated predictive estimates for Bayesian hierarchical regression models (BHRMs). BHRMs are popular for modeling complex dependence structures (e.g., Gaussian processes and Gaussian Markov random fields) but can be computationally expensive to run. Cross-validation (CV) is, therefore, not a common practice to evaluate the predictive performance of BHRMs. Our method circumvents the need to rerun computationally costly estimation methods for each cross-validation fold and makes CV more feasible for large BHRMs. We shift the CV problem from probability-based sampling to a familiar and straightforward optimization problem by conditioning on the variance-covariance parameters. Our approximation applies to leave-one-out CV and leave-one-cluster-out CV, the latter of which is more appropriate for models with complex dependencies. In many cases, this produces estimates equivalent to full CV. We provide theoretical results, demonstrate the efficacy of our method on publicly available data and in simulations, and compare the model performance with several competing methods for CV approximation. Code and other supplementary materials available online.
UR - http://www.scopus.com/inward/record.url?scp=85209577350&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85209577350&partnerID=8YFLogxK
U2 - 10.1080/10618600.2024.2404711
DO - 10.1080/10618600.2024.2404711
M3 - Article
AN - SCOPUS:85209577350
SN - 1061-8600
JO - Journal of Computational and Graphical Statistics
JF - Journal of Computational and Graphical Statistics
ER -