{"title":"Semi Supervised Learning for Human Activity Recognition Using Depth Cameras","authors":"Moustafa F. Mabrouk, Nagia M. Ghanem, M. Ismail","doi":"10.1109/ICMLA.2015.170","DOIUrl":null,"url":null,"abstract":"Human action recognition is a very active research topic in computer vision and pattern recognition. Recently, it has shown a great potential for human action recognition using the 3D depth data captured by the promising RGB-D cameras, and particularly, the Microsoft Kinect which has made high resolution real-time depth cheaply available. Several features and descriptors have been proposed for depth based action recognition, and they have given high results when recognizing the actions, but one dilemma always exists, the labeled data given, which are manually set by humans. They are not enough to build the system, especially that the use of human action recognition is mainly for surveillance of people activities. In this paper, the paucity of labeled data is addressed, by the popular semi supervision machine learning technique \"co-training\", which makes full use of unlabeled samples of two different independent views. Through the experiments on two popular datasets (MSR Action 3D, and MSR DailyActivity 3D), we demonstrate that our proposed framework outperforms the state of art. It improves the accuracy up to 83% in case of MSR Action 3D, and up to 80% MSR DailyActivity 3D, using the same number of labeled samples.","PeriodicalId":288427,"journal":{"name":"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2015.170","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Human action recognition is a very active research topic in computer vision and pattern recognition. Recently, it has shown a great potential for human action recognition using the 3D depth data captured by the promising RGB-D cameras, and particularly, the Microsoft Kinect which has made high resolution real-time depth cheaply available. Several features and descriptors have been proposed for depth based action recognition, and they have given high results when recognizing the actions, but one dilemma always exists, the labeled data given, which are manually set by humans. They are not enough to build the system, especially that the use of human action recognition is mainly for surveillance of people activities. In this paper, the paucity of labeled data is addressed, by the popular semi supervision machine learning technique "co-training", which makes full use of unlabeled samples of two different independent views. Through the experiments on two popular datasets (MSR Action 3D, and MSR DailyActivity 3D), we demonstrate that our proposed framework outperforms the state of art. It improves the accuracy up to 83% in case of MSR Action 3D, and up to 80% MSR DailyActivity 3D, using the same number of labeled samples.