TY - JOUR
T1 - Communication-Efficient Distributed Statistical Inference
AU - Jordan, Michael I.
AU - Lee, Jason D.
AU - Yang, Yun
N1 - Publisher Copyright:
© 2018, © 2018 American Statistical Association.
Copyright:
Copyright 2019 Elsevier B.V., All rights reserved.
PY - 2019/4/3
Y1 - 2019/4/3
N2 - We present a communication-efficient surrogate likelihood (CSL) framework for solving distributed statistical inference problems. CSL provides a communication-efficient surrogate to the global likelihood that can be used for low-dimensional estimation, high-dimensional regularized estimation, and Bayesian inference. For low-dimensional estimation, CSL provably improves upon naive averaging schemes and facilitates the construction of confidence intervals. For high-dimensional regularized estimation, CSL leads to a minimax-optimal estimator with controlled communication cost. For Bayesian inference, CSL can be used to form a communication-efficient quasi-posterior distribution that converges to the true posterior. This quasi-posterior procedure significantly improves the computational efficiency of Markov chain Monte Carlo (MCMC) algorithms even in a nondistributed setting. We present both theoretical analysis and experiments to explore the properties of the CSL approximation. Supplementary materials for this article are available online.
AB - We present a communication-efficient surrogate likelihood (CSL) framework for solving distributed statistical inference problems. CSL provides a communication-efficient surrogate to the global likelihood that can be used for low-dimensional estimation, high-dimensional regularized estimation, and Bayesian inference. For low-dimensional estimation, CSL provably improves upon naive averaging schemes and facilitates the construction of confidence intervals. For high-dimensional regularized estimation, CSL leads to a minimax-optimal estimator with controlled communication cost. For Bayesian inference, CSL can be used to form a communication-efficient quasi-posterior distribution that converges to the true posterior. This quasi-posterior procedure significantly improves the computational efficiency of Markov chain Monte Carlo (MCMC) algorithms even in a nondistributed setting. We present both theoretical analysis and experiments to explore the properties of the CSL approximation. Supplementary materials for this article are available online.
KW - Communication efficiency
KW - Distributed inference
KW - Likelihood approximation
UR - http://www.scopus.com/inward/record.url?scp=85051502370&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85051502370&partnerID=8YFLogxK
U2 - 10.1080/01621459.2018.1429274
DO - 10.1080/01621459.2018.1429274
M3 - Article
SN - 0162-1459
VL - 114
SP - 668
EP - 681
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 526
ER -