TY - JOUR
T1 - Mixed-membership naive bayes models
AU - Shan, Hanhuai
AU - Banerjee, Arindam
N1 - Funding Information:
Acknowledgements We want to warmly thank Nikunj Oza for valuable input on discriminative mixed membership models. The research was supported by National Science Foundation grants IIS-0812183, IIS-0916750, National Science Foundation CAREER grant IIS-0953274, and National Aeronautics and Space Administration grant NNX08AC36A.
PY - 2011/7
Y1 - 2011/7
N2 - In recent years, mixture models have found widespread usage in discovering latent cluster structure from data. A popular special case of finite mixturemodels is the family of naive Bayes (NB) models, where the probability of a feature vector factorizes over the features for any given component of the mixture. Despite their popularity, naive Bayes models do not allow data points to belong to different component clusters with varying degrees, i.e., mixed memberships, which puts a restriction on their modeling ability. In this paper, we propose mixed-membership naïve Bayes (MMNB) models. On one hand, MMNB can be viewed as a generalization of NB by putting a Dirichlet prior on top to allow mixed memberships. On the other hand, MMNB can also be viewed as a generalization of latent Dirichlet allocation (LDA) with the ability to handle heterogeneous feature vectors with different types of features, e.g., real, categorical, etc.. We propose two variational inference algorithms to learnMMNBmodels. The first one is based on ideas originally used in LDA, and the second one uses substantially fewer variational parameters, leading to a significantly faster algorithm. Further, we extend MMNB/LDA to discriminative mixed-membership models for classification by suitably combining MMNB/LDA with multi-class logistic regression. The efficacy of the proposed mixed-membership models is demonstrated by extensive experiments on several datasets, including UCI benchmarks, recommendation systems, and text datasets.
AB - In recent years, mixture models have found widespread usage in discovering latent cluster structure from data. A popular special case of finite mixturemodels is the family of naive Bayes (NB) models, where the probability of a feature vector factorizes over the features for any given component of the mixture. Despite their popularity, naive Bayes models do not allow data points to belong to different component clusters with varying degrees, i.e., mixed memberships, which puts a restriction on their modeling ability. In this paper, we propose mixed-membership naïve Bayes (MMNB) models. On one hand, MMNB can be viewed as a generalization of NB by putting a Dirichlet prior on top to allow mixed memberships. On the other hand, MMNB can also be viewed as a generalization of latent Dirichlet allocation (LDA) with the ability to handle heterogeneous feature vectors with different types of features, e.g., real, categorical, etc.. We propose two variational inference algorithms to learnMMNBmodels. The first one is based on ideas originally used in LDA, and the second one uses substantially fewer variational parameters, leading to a significantly faster algorithm. Further, we extend MMNB/LDA to discriminative mixed-membership models for classification by suitably combining MMNB/LDA with multi-class logistic regression. The efficacy of the proposed mixed-membership models is demonstrated by extensive experiments on several datasets, including UCI benchmarks, recommendation systems, and text datasets.
KW - Generative models
KW - Latent Dirichlet allocation
KW - Logistic regression
KW - Mixed-membership
KW - Naive Bayes
KW - Variational inference
UR - http://www.scopus.com/inward/record.url?scp=79960097163&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79960097163&partnerID=8YFLogxK
U2 - 10.1007/s10618-010-0198-2
DO - 10.1007/s10618-010-0198-2
M3 - Article
AN - SCOPUS:79960097163
SN - 1384-5810
VL - 23
SP - 1
EP - 62
JO - Data Mining and Knowledge Discovery
JF - Data Mining and Knowledge Discovery
IS - 1
ER -