TY - GEN
T1 - Multi-modal social interaction recognition using view-invariant features
AU - Trabelsi, Rim
AU - Varadarajan, Jagannadan
AU - Pei, Yong
AU - Zhang, Le
AU - Jabri, Issam
AU - Bouallegue, Ammar
AU - Moulin, Pierre
N1 - Publisher Copyright:
© 2017 Copyright is held by the owner/author(s).
PY - 2017/11/13
Y1 - 2017/11/13
N2 - This paper addresses the issue of analyzing social interactions between humans in videos. We focus on recognizing dyadic human interactions through multi-modal data, specifically, depth, color and skeleton sequences. Firstly, we introduce a new person-centric proxemic descriptor, named PROF, extracted from skeleton data able to incorporate intrinsic and extrinsic distances between two interacting persons in a view-variant scheme. Then, a novel key frame selection approach is introduced to identify salient instants of the interaction sequence based on the joint energy. From RGBD videos, more holistic CNN features are extracted by applying an adaptive pre-trained CNNs on optical flow frames. Features from three modalities are combined then classified using linear SVM. Finally, extensive experiments have been carried on two multi-modal and multi-view interactions datasets prove the robustness of the introduced approach comparing to state-of-the-art methods.
AB - This paper addresses the issue of analyzing social interactions between humans in videos. We focus on recognizing dyadic human interactions through multi-modal data, specifically, depth, color and skeleton sequences. Firstly, we introduce a new person-centric proxemic descriptor, named PROF, extracted from skeleton data able to incorporate intrinsic and extrinsic distances between two interacting persons in a view-variant scheme. Then, a novel key frame selection approach is introduced to identify salient instants of the interaction sequence based on the joint energy. From RGBD videos, more holistic CNN features are extracted by applying an adaptive pre-trained CNNs on optical flow frames. Features from three modalities are combined then classified using linear SVM. Finally, extensive experiments have been carried on two multi-modal and multi-view interactions datasets prove the robustness of the introduced approach comparing to state-of-the-art methods.
KW - Multi-modal data
KW - Multi-view
KW - Social interaction
UR - http://www.scopus.com/inward/record.url?scp=85041189722&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85041189722&partnerID=8YFLogxK
U2 - 10.1145/3139491.3139501
DO - 10.1145/3139491.3139501
M3 - Conference contribution
AN - SCOPUS:85041189722
T3 - ISIAA 2017 - Proceedings of the 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents, Co-located with ICMI 2017
SP - 47
EP - 48
BT - ISIAA 2017 - Proceedings of the 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents, Co-located with ICMI 2017
A2 - Lefevre, Fabrice
A2 - Chaminade, Thierry
A2 - Ngyuen, Noel
A2 - Ochs, Magalie
PB - Association for Computing Machinery, Inc
T2 - 1st ACM SIGCHI International Workshop on Investigating Social Interactions with Artificial Agents, ISIAA 2017
Y2 - 13 November 2017
ER -