TY - GEN
T1 - Learning a task-specific deep architecture for clustering
AU - Wang, Zhangyang
AU - Chang, Shiyu
AU - Zhou, Jiayu
AU - Wang, Meng
AU - Huang, Thomas S.
N1 - Publisher Copyright:
Copyright © by SIAM.
PY - 2016
Y1 - 2016
N2 - While sparse coding-based clustering methods have shown to be successful, their bottlenecks in both efficiency and scalability limit the practical usage. In recent years, deep learning has been proved to be a highly effective, efficient and scalable feature learning tool. In this paper, we propose to emulate the sparse coding-based clustering pipeline in the context of deep learning, leading to a carefully crafted deep model benefiting from both. A feed-forward network structure, named TAGnet, is constructed based on a graph-regularized sparse coding algorithm. It is then trained with task-specific loss functions from end to end. We discover that connecting deep learning to sparse coding benefits not only the model performance, but also its initialization and interpretation. Moreover, by introducing auxiliary clustering tasks to the intermediate feature hierarchy, we formulate DTAGnet and obtain a further performance boost. Extensive experiments demonstrate that the proposed model gains remarkable margins over several state-of-the-art methods.
AB - While sparse coding-based clustering methods have shown to be successful, their bottlenecks in both efficiency and scalability limit the practical usage. In recent years, deep learning has been proved to be a highly effective, efficient and scalable feature learning tool. In this paper, we propose to emulate the sparse coding-based clustering pipeline in the context of deep learning, leading to a carefully crafted deep model benefiting from both. A feed-forward network structure, named TAGnet, is constructed based on a graph-regularized sparse coding algorithm. It is then trained with task-specific loss functions from end to end. We discover that connecting deep learning to sparse coding benefits not only the model performance, but also its initialization and interpretation. Moreover, by introducing auxiliary clustering tasks to the intermediate feature hierarchy, we formulate DTAGnet and obtain a further performance boost. Extensive experiments demonstrate that the proposed model gains remarkable margins over several state-of-the-art methods.
UR - http://www.scopus.com/inward/record.url?scp=84991596493&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84991596493&partnerID=8YFLogxK
U2 - 10.1137/1.9781611974348.42
DO - 10.1137/1.9781611974348.42
M3 - Conference contribution
AN - SCOPUS:84991596493
T3 - 16th SIAM International Conference on Data Mining 2016, SDM 2016
SP - 369
EP - 377
BT - 16th SIAM International Conference on Data Mining 2016, SDM 2016
A2 - Venkatasubramanian, Sanjay Chawla
A2 - Meira, Wagner
PB - Society for Industrial and Applied Mathematics Publications
T2 - 16th SIAM International Conference on Data Mining 2016, SDM 2016
Y2 - 5 May 2016 through 7 May 2016
ER -