@inproceedings{5635789c87914300b5cee65ed9ce91cd,
title = "GPU accelerated probabilistic latent sequential motifs for activity analysis",
abstract = "In this paper, we present an optimized GPU based implementation of Probabilistic Latent Sequential motifs (PLSM) that was proposed for sequential pattern mining from video sequences. PLSM mines for recurrent sequential patterns from documents given as word-time occurrences, and outputs a set of sequential activity motifs and their starting occurrences. PLSM{\textquoteright}s uniqueness comes from modeling the co-occurrence and temporal order in which the words occur within a temporal window while also dealing with activities which occur concurrently in the video. However, the expectation-maximization algorithm used in PLSM has a very high time complexity due to complex nested loops, requiring several dimensionality reduction steps before invoking PLSM. In order to truly realize the benefits of the model, we propose two GPU based implementations of PLSM called GPU-pLSM (sparse and dense). The two implementations differ based on whether the entire word-count matrix (dense) or only the non-zero entries (sparse) are considered in inferring the latent motifs respectively. Our implementation achieves an impressive 265X and 366X times speed up for dense and sparse approaches respectively on NVIDIA GeForce GTX Titan. This speed up enables us to remove several pre-processing and dimension reduction steps used to generate the input temporal documents and thus apply PLSM directly on the input documents. We validate our results through qualitative comparisons of the inferred motifs on two different publicly available datasets. Quantitative comparison on document reconstruction based abnormality measure show that both GPU-PLSM and PLSA+PLSM are strongly correlated.",
keywords = "Activity analysis, CUDA, GPU, Motifs, PLSA, PLSM, Topic models",
author = "Mohiuddin, {Khaja Wasif} and Jagannadan Varadarajan and R{\'e}mi Emonet and Odobez, {Jean Marc} and Pierre Moulin",
note = "Funding Information: This study is supported by the research grant for the Human Centered Cyber-physical Systems Programme at the Advanced Digital Sciences Center from Singapore's Agency for Science, Technology and Research (A*STAR). The authors thank NVIDIA for donating GPUs to support this research work. Funding Information: This study is supported by the research grant for the Human Centered Cyber-physical Systems Pro-gramme at the Advanced Digital Sciences Center from Singapore's Agency for Science, Technology and Research (A*STAR). The authors thank NVIDIA for donating GPUs to support this research work. Publisher Copyright: {\textcopyright} 2018 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved.; 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, VISIGRAPP 2018 ; Conference date: 27-01-2018 Through 29-01-2018",
year = "2018",
doi = "10.5220/0006537904090418",
language = "English (US)",
series = "VISIGRAPP 2018 - Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications",
publisher = "SciTePress",
pages = "409--418",
editor = "Francisco Imai and Alain Tremeau and Jose Braz",
booktitle = "VISAPP",
}