TY - JOUR
T1 - Skinny Gibbs
T2 - A Consistent and Scalable Gibbs Sampler for Model Selection
AU - Narisetty, Naveen N.
AU - Shen, Juan
AU - He, Xuming
N1 - Funding Information:
The authors thank Professor Faming Liang for providing them the code to perform model selection based on the Bayesian Subset Regression. The research is partially supported by the NSF Awards DMS-1307566, DMS-1607840, DMS-1811768, and the Chinese National Natural Science Projects 11129101, 11501123, and 11690012.
Publisher Copyright:
© 2018, © 2018 American Statistical Association.
PY - 2019/7/3
Y1 - 2019/7/3
N2 - We consider the computational and statistical issues for high-dimensional Bayesian model selection under the Gaussian spike and slab priors. To avoid large matrix computations needed in a standard Gibbs sampler, we propose a novel Gibbs sampler called “Skinny Gibbs” which is much more scalable to high-dimensional problems, both in memory and in computational efficiency. In particular, its computational complexity grows only linearly in p, the number of predictors, while retaining the property of strong model selection consistency even when p is much greater than the sample size n. The present article focuses on logistic regression due to its broad applicability as a representative member of the generalized linear models. We compare our proposed method with several leading variable selection methods through a simulation study to show that Skinny Gibbs has a strong performance as indicated by our theoretical work. Supplementary materials for this article are available online.
AB - We consider the computational and statistical issues for high-dimensional Bayesian model selection under the Gaussian spike and slab priors. To avoid large matrix computations needed in a standard Gibbs sampler, we propose a novel Gibbs sampler called “Skinny Gibbs” which is much more scalable to high-dimensional problems, both in memory and in computational efficiency. In particular, its computational complexity grows only linearly in p, the number of predictors, while retaining the property of strong model selection consistency even when p is much greater than the sample size n. The present article focuses on logistic regression due to its broad applicability as a representative member of the generalized linear models. We compare our proposed method with several leading variable selection methods through a simulation study to show that Skinny Gibbs has a strong performance as indicated by our theoretical work. Supplementary materials for this article are available online.
KW - Bayesian computation
KW - Gibbs sampling
KW - High-dimensional data
KW - Logistic regression
KW - Scalable computation
KW - Variable selection
UR - http://www.scopus.com/inward/record.url?scp=85052082791&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85052082791&partnerID=8YFLogxK
U2 - 10.1080/01621459.2018.1482754
DO - 10.1080/01621459.2018.1482754
M3 - Article
AN - SCOPUS:85052082791
SN - 0162-1459
VL - 114
SP - 1205
EP - 1217
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 527
ER -