Robust multi-modal cues for dyadic human interaction recognition

Rim Trabelsi, Jagannadan Varadarajan, Yong Pei, Le Zhang, Issam Jabri, Ammar Bouallegue, Pierre Moulin

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Activity analysis methods usually tend to focus on elementary human actions but ignore to analyze complex scenarios. In this paper, we focus particularly on classifying interactions between two persons in a supervised fashion. We propose a robust multi-modal proxemic descriptor based on 3D joint locations, depth and color videos. The proposed descriptor incorporates inter-person and intraperson joint distances calculated from 3D skeleton data and multiframe dense optical flow features obtained from the application of temporal Convolutional neural networks (CNN) on depth and color images. The descriptors from the three modalities are derived from sparse key-frames surrounding high activity content and fused using a linear SVM classifier. Through experiments on two publicly available RGB-D interaction datasets, we show that our method can efficiently classify complex interactions using only short video snippet, outperforming existing state-of-the-art results.

Original languageEnglish (US)
Title of host publicationMUSA2 2017 - Proceedings of the Workshop on Multimodal Understanding of Social, Affective and Subjective Attributes, co-located with MM 2017
PublisherAssociation for Computing Machinery
Pages47-53
Number of pages7
ISBN (Electronic)9781450355094
DOIs
StatePublished - Oct 27 2017
Event1st ACM MM Workshop on Multimodal Understanding of Social, Affective and Subjective Attributes, MUSA2 2017 - Mountain View, United States
Duration: Oct 27 2017 → …

Publication series

NameMUSA2 2017 - Proceedings of the Workshop on Multimodal Understanding of Social, Affective and Subjective Attributes, co-located with MM 2017

Other

Other1st ACM MM Workshop on Multimodal Understanding of Social, Affective and Subjective Attributes, MUSA2 2017
Country/TerritoryUnited States
CityMountain View
Period10/27/17 → …

Keywords

  • CNN features
  • Interaction recognition
  • Multi-modal features
  • RGB-D
  • Skeleton data

ASJC Scopus subject areas

  • Human-Computer Interaction
  • Computer Graphics and Computer-Aided Design

Fingerprint

Dive into the research topics of 'Robust multi-modal cues for dyadic human interaction recognition'. Together they form a unique fingerprint.

Cite this