{"title":"基于视点不变3D驾驶员身体姿势的活动识别","authors":"Manuel Martin, D. Lerch, M. Voit","doi":"10.1109/IV55152.2023.10186682","DOIUrl":null,"url":null,"abstract":"Driver monitoring will be required in many countries for all new vehicles with automation functions. While the common approach for this task is face and eye gaze monitoring, cameras with a wider field of view have already been introduced in some cars. However, their mounting position can change between vehicle models. To minimize data collection efforts and to facilitate data reuse it is important for algorithms to be able to deal with a changing environment. We conduct an experiment comparing the performance of video-based models with 3D body pose-based activity recognition methods with regards to sensor and position changes. We introduce a modular activity recognition pipeline which uses a sensor independent representation including the 3D body pose of the driver, 3D dynamic object positions and 3D interior positions. We show that while video-based models offer the best quality overall, when trained and tested on the same camera view, body pose-based methods can be far more robust to positional changes. Moreover, augmentation reduces the performance drop across views to 35% compared to 83% without augmentation and 90% for video-based models.","PeriodicalId":195148,"journal":{"name":"2023 IEEE Intelligent Vehicles Symposium (IV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Viewpoint Invariant 3D Driver Body Pose-Based Activity Recognition\",\"authors\":\"Manuel Martin, D. Lerch, M. Voit\",\"doi\":\"10.1109/IV55152.2023.10186682\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Driver monitoring will be required in many countries for all new vehicles with automation functions. While the common approach for this task is face and eye gaze monitoring, cameras with a wider field of view have already been introduced in some cars. However, their mounting position can change between vehicle models. To minimize data collection efforts and to facilitate data reuse it is important for algorithms to be able to deal with a changing environment. We conduct an experiment comparing the performance of video-based models with 3D body pose-based activity recognition methods with regards to sensor and position changes. We introduce a modular activity recognition pipeline which uses a sensor independent representation including the 3D body pose of the driver, 3D dynamic object positions and 3D interior positions. We show that while video-based models offer the best quality overall, when trained and tested on the same camera view, body pose-based methods can be far more robust to positional changes. Moreover, augmentation reduces the performance drop across views to 35% compared to 83% without augmentation and 90% for video-based models.\",\"PeriodicalId\":195148,\"journal\":{\"name\":\"2023 IEEE Intelligent Vehicles Symposium (IV)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE Intelligent Vehicles Symposium (IV)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IV55152.2023.10186682\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE Intelligent Vehicles Symposium (IV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IV55152.2023.10186682","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Viewpoint Invariant 3D Driver Body Pose-Based Activity Recognition
Driver monitoring will be required in many countries for all new vehicles with automation functions. While the common approach for this task is face and eye gaze monitoring, cameras with a wider field of view have already been introduced in some cars. However, their mounting position can change between vehicle models. To minimize data collection efforts and to facilitate data reuse it is important for algorithms to be able to deal with a changing environment. We conduct an experiment comparing the performance of video-based models with 3D body pose-based activity recognition methods with regards to sensor and position changes. We introduce a modular activity recognition pipeline which uses a sensor independent representation including the 3D body pose of the driver, 3D dynamic object positions and 3D interior positions. We show that while video-based models offer the best quality overall, when trained and tested on the same camera view, body pose-based methods can be far more robust to positional changes. Moreover, augmentation reduces the performance drop across views to 35% compared to 83% without augmentation and 90% for video-based models.