TY - GEN
T1 - Beta process multiple kernel learning
AU - Ni, Bingbing
AU - Li, Teng
AU - Moulin, Pierre
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2014/9/24
Y1 - 2014/9/24
N2 - In kernel based learning, the kernel trick transforms the original representation of a feature instance into a vector of similarities with the training feature instances, known as kernel representation. However, feature instances are sometimes ambiguous and the kernel representation calculated based on them do not possess any discriminative information, which can eventually harm the trained classifier. To address this issue, we propose to automatically select good feature instances when calculating the kernel representation in multiple kernel learning. Specifically, for the kernel representation calculated for each input feature instance, we multiply it element-wise with a latent binary vector named as instance selection variables, which targets at selecting good instances and attenuate the effect of ambiguous ones in the resulting new kernel representation. Beta process is employed for generating the prior distribution for the latent instance selection variables. We then propose a Bayesian graphical model which integrates both MKL learning and inference for the distribution of the latent instance selection variables. Variational inference is derived for model learning under a max-margin principle. Our method is called Beta process multiple kernel learning. Extensive experiments demonstrate the effectiveness of our method on instance selection and its high discriminative capability for various classification problems in vision.
AB - In kernel based learning, the kernel trick transforms the original representation of a feature instance into a vector of similarities with the training feature instances, known as kernel representation. However, feature instances are sometimes ambiguous and the kernel representation calculated based on them do not possess any discriminative information, which can eventually harm the trained classifier. To address this issue, we propose to automatically select good feature instances when calculating the kernel representation in multiple kernel learning. Specifically, for the kernel representation calculated for each input feature instance, we multiply it element-wise with a latent binary vector named as instance selection variables, which targets at selecting good instances and attenuate the effect of ambiguous ones in the resulting new kernel representation. Beta process is employed for generating the prior distribution for the latent instance selection variables. We then propose a Bayesian graphical model which integrates both MKL learning and inference for the distribution of the latent instance selection variables. Variational inference is derived for model learning under a max-margin principle. Our method is called Beta process multiple kernel learning. Extensive experiments demonstrate the effectiveness of our method on instance selection and its high discriminative capability for various classification problems in vision.
KW - Beta process
KW - classification
KW - instance selection
KW - multiple kernel learning
KW - variational inference
UR - http://www.scopus.com/inward/record.url?scp=84911388986&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84911388986&partnerID=8YFLogxK
U2 - 10.1109/CVPR.2014.128
DO - 10.1109/CVPR.2014.128
M3 - Conference contribution
AN - SCOPUS:84911388986
T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
SP - 963
EP - 970
BT - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
PB - IEEE Computer Society
T2 - 27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014
Y2 - 23 June 2014 through 28 June 2014
ER -