R. Alazrai, Mohammad Hababeh, B. Alsaify, M. Daoud
{"title":"Anatomical Planes-Based Representation for Recognizing Two-Person Interactions from Partially Observed Video Sequences: A Feasibility Study","authors":"R. Alazrai, Mohammad Hababeh, B. Alsaify, M. Daoud","doi":"10.1109/ICEEE52452.2021.9415910","DOIUrl":null,"url":null,"abstract":"This paper presents a new approach for two-person interaction recognition from partially observed video sequences. The proposed approach employs the 3D joint positions captured by a Microsoft Kinect sensor to construct a view-invariant anatomical planes-based descriptor, called the two-person motion-pose geometric descriptor (TP-MPGD), that quantifies the activities performed by two interacting persons at each video frame. Using the TP-MPGDs extracted from the frames of the input videos, we construct a two-phase classification framework to recognize the class of the interaction performed by two persons. The performance of the proposed approach has been evaluated using a publicly available interaction dataset that comprises the 3D joint positions data recorded using the Kinect sensor for 21 pairs of subjects while performing eight interactions. Moreover, we have developed five different evaluation scenarios, including one evaluation scenario that is based on fully observed video sequences and four other evaluation scenarios that are based on partially observed video sequences. The classification accuracies obtained for each of the five evaluation scenarios demonstrate the feasibility of our proposed approach to recognize two-person interactions from fully observed and partially observed video sequences.","PeriodicalId":429645,"journal":{"name":"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEEE52452.2021.9415910","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents a new approach for two-person interaction recognition from partially observed video sequences. The proposed approach employs the 3D joint positions captured by a Microsoft Kinect sensor to construct a view-invariant anatomical planes-based descriptor, called the two-person motion-pose geometric descriptor (TP-MPGD), that quantifies the activities performed by two interacting persons at each video frame. Using the TP-MPGDs extracted from the frames of the input videos, we construct a two-phase classification framework to recognize the class of the interaction performed by two persons. The performance of the proposed approach has been evaluated using a publicly available interaction dataset that comprises the 3D joint positions data recorded using the Kinect sensor for 21 pairs of subjects while performing eight interactions. Moreover, we have developed five different evaluation scenarios, including one evaluation scenario that is based on fully observed video sequences and four other evaluation scenarios that are based on partially observed video sequences. The classification accuracies obtained for each of the five evaluation scenarios demonstrate the feasibility of our proposed approach to recognize two-person interactions from fully observed and partially observed video sequences.