基于深度学习的图像处理人类动作识别

A. P. Ismail, Muhammad Afiq Bin Azahar, N. Tahir, K. Daud, Nazirah Mohamat Kasim
{"title":"基于深度学习的图像处理人类动作识别","authors":"A. P. Ismail, Muhammad Afiq Bin Azahar, N. Tahir, K. Daud, Nazirah Mohamat Kasim","doi":"10.1109/ICCSCE58721.2023.10237158","DOIUrl":null,"url":null,"abstract":"The advancement of artificial intelligence (AI) has bought many advances to human society as a whole. By using daily activities and integrating the technology from the fruits of AI, we can manage to gain further access to knowledge we can only begin to imagine. In identifying human action recognition (HAR); processing photos and videos to discern whether a human is present, then mapping the subject classified, which lastly determines the action being carried out is the objective. To achieve this, various steps are taken and careful approach is required, with the extensive amount of research, numerous troubleshooting and experimentation is required. The AI architecture has to learn from dataset collected for it to discern the identification of action properly. HAR is achieved by using Python code using real-time webcam feed. Human pose detection library known as MediaPipe Pose Detection detects human anatomy from input through joints key-points. MediaPipe algorithm that extract features in x-y-z axis with visibility (four variables) and the extracted data is trained using CNN-LSTM based on the trained and tested algorithm classifier model. The output obtained produced an RGB-skeleton and an action label on the detected subject as standing, waving, walking and sitting, has yielded good results.","PeriodicalId":287947,"journal":{"name":"2023 IEEE 13th International Conference on Control System, Computing and Engineering (ICCSCE)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Human Action Recognition (HAR) using Image Processing on Deep Learning\",\"authors\":\"A. P. Ismail, Muhammad Afiq Bin Azahar, N. Tahir, K. Daud, Nazirah Mohamat Kasim\",\"doi\":\"10.1109/ICCSCE58721.2023.10237158\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The advancement of artificial intelligence (AI) has bought many advances to human society as a whole. By using daily activities and integrating the technology from the fruits of AI, we can manage to gain further access to knowledge we can only begin to imagine. In identifying human action recognition (HAR); processing photos and videos to discern whether a human is present, then mapping the subject classified, which lastly determines the action being carried out is the objective. To achieve this, various steps are taken and careful approach is required, with the extensive amount of research, numerous troubleshooting and experimentation is required. The AI architecture has to learn from dataset collected for it to discern the identification of action properly. HAR is achieved by using Python code using real-time webcam feed. Human pose detection library known as MediaPipe Pose Detection detects human anatomy from input through joints key-points. MediaPipe algorithm that extract features in x-y-z axis with visibility (four variables) and the extracted data is trained using CNN-LSTM based on the trained and tested algorithm classifier model. The output obtained produced an RGB-skeleton and an action label on the detected subject as standing, waving, walking and sitting, has yielded good results.\",\"PeriodicalId\":287947,\"journal\":{\"name\":\"2023 IEEE 13th International Conference on Control System, Computing and Engineering (ICCSCE)\",\"volume\":\"37 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-08-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE 13th International Conference on Control System, Computing and Engineering (ICCSCE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCSCE58721.2023.10237158\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 13th International Conference on Control System, Computing and Engineering (ICCSCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCSCE58721.2023.10237158","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

人工智能(AI)的进步给整个人类社会带来了许多进步。通过使用日常活动和整合人工智能成果的技术,我们可以设法获得更多的知识,我们只能开始想象。在识别人类行为识别(HAR);处理照片和视频来辨别是否有人在场,然后绘制分类的主题,最后确定正在进行的行动是目标。为了实现这一目标,需要采取各种步骤和谨慎的方法,需要进行大量的研究、大量的故障排除和实验。人工智能架构必须从收集的数据集中学习,以便正确识别动作。HAR是通过使用Python代码使用实时网络摄像头feed实现的。人体姿态检测库称为MediaPipe姿态检测检测人体解剖从输入通过关节关键点。MediaPipe算法提取x-y-z轴上具有可见性的特征(四个变量),提取的数据在训练和测试的算法分类器模型的基础上,使用CNN-LSTM进行训练。所得到的输出产生了rgb骨架和被检测对象的站立、摆动、行走和坐姿动作标签,取得了良好的效果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Human Action Recognition (HAR) using Image Processing on Deep Learning
The advancement of artificial intelligence (AI) has bought many advances to human society as a whole. By using daily activities and integrating the technology from the fruits of AI, we can manage to gain further access to knowledge we can only begin to imagine. In identifying human action recognition (HAR); processing photos and videos to discern whether a human is present, then mapping the subject classified, which lastly determines the action being carried out is the objective. To achieve this, various steps are taken and careful approach is required, with the extensive amount of research, numerous troubleshooting and experimentation is required. The AI architecture has to learn from dataset collected for it to discern the identification of action properly. HAR is achieved by using Python code using real-time webcam feed. Human pose detection library known as MediaPipe Pose Detection detects human anatomy from input through joints key-points. MediaPipe algorithm that extract features in x-y-z axis with visibility (four variables) and the extracted data is trained using CNN-LSTM based on the trained and tested algorithm classifier model. The output obtained produced an RGB-skeleton and an action label on the detected subject as standing, waving, walking and sitting, has yielded good results.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信