R. Alazrai, Mohammad Hababeh, B. Alsaify, M. Daoud
{"title":"基于解剖平面的表征从部分观察到的视频序列中识别两人互动:可行性研究","authors":"R. Alazrai, Mohammad Hababeh, B. Alsaify, M. Daoud","doi":"10.1109/ICEEE52452.2021.9415910","DOIUrl":null,"url":null,"abstract":"This paper presents a new approach for two-person interaction recognition from partially observed video sequences. The proposed approach employs the 3D joint positions captured by a Microsoft Kinect sensor to construct a view-invariant anatomical planes-based descriptor, called the two-person motion-pose geometric descriptor (TP-MPGD), that quantifies the activities performed by two interacting persons at each video frame. Using the TP-MPGDs extracted from the frames of the input videos, we construct a two-phase classification framework to recognize the class of the interaction performed by two persons. The performance of the proposed approach has been evaluated using a publicly available interaction dataset that comprises the 3D joint positions data recorded using the Kinect sensor for 21 pairs of subjects while performing eight interactions. Moreover, we have developed five different evaluation scenarios, including one evaluation scenario that is based on fully observed video sequences and four other evaluation scenarios that are based on partially observed video sequences. The classification accuracies obtained for each of the five evaluation scenarios demonstrate the feasibility of our proposed approach to recognize two-person interactions from fully observed and partially observed video sequences.","PeriodicalId":429645,"journal":{"name":"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Anatomical Planes-Based Representation for Recognizing Two-Person Interactions from Partially Observed Video Sequences: A Feasibility Study\",\"authors\":\"R. Alazrai, Mohammad Hababeh, B. Alsaify, M. Daoud\",\"doi\":\"10.1109/ICEEE52452.2021.9415910\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a new approach for two-person interaction recognition from partially observed video sequences. The proposed approach employs the 3D joint positions captured by a Microsoft Kinect sensor to construct a view-invariant anatomical planes-based descriptor, called the two-person motion-pose geometric descriptor (TP-MPGD), that quantifies the activities performed by two interacting persons at each video frame. Using the TP-MPGDs extracted from the frames of the input videos, we construct a two-phase classification framework to recognize the class of the interaction performed by two persons. The performance of the proposed approach has been evaluated using a publicly available interaction dataset that comprises the 3D joint positions data recorded using the Kinect sensor for 21 pairs of subjects while performing eight interactions. Moreover, we have developed five different evaluation scenarios, including one evaluation scenario that is based on fully observed video sequences and four other evaluation scenarios that are based on partially observed video sequences. The classification accuracies obtained for each of the five evaluation scenarios demonstrate the feasibility of our proposed approach to recognize two-person interactions from fully observed and partially observed video sequences.\",\"PeriodicalId\":429645,\"journal\":{\"name\":\"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)\",\"volume\":\"93 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-04-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICEEE52452.2021.9415910\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 8th International Conference on Electrical and Electronics Engineering (ICEEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEEE52452.2021.9415910","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Anatomical Planes-Based Representation for Recognizing Two-Person Interactions from Partially Observed Video Sequences: A Feasibility Study
This paper presents a new approach for two-person interaction recognition from partially observed video sequences. The proposed approach employs the 3D joint positions captured by a Microsoft Kinect sensor to construct a view-invariant anatomical planes-based descriptor, called the two-person motion-pose geometric descriptor (TP-MPGD), that quantifies the activities performed by two interacting persons at each video frame. Using the TP-MPGDs extracted from the frames of the input videos, we construct a two-phase classification framework to recognize the class of the interaction performed by two persons. The performance of the proposed approach has been evaluated using a publicly available interaction dataset that comprises the 3D joint positions data recorded using the Kinect sensor for 21 pairs of subjects while performing eight interactions. Moreover, we have developed five different evaluation scenarios, including one evaluation scenario that is based on fully observed video sequences and four other evaluation scenarios that are based on partially observed video sequences. The classification accuracies obtained for each of the five evaluation scenarios demonstrate the feasibility of our proposed approach to recognize two-person interactions from fully observed and partially observed video sequences.