S. Sreenath, D. Daniels, Apparaju S. D. Ganesh, Yashaswi S. Kuruganti, Rajeevlochana G. Chittawadigi
{"title":"Monocular Tracking of Human Hand on a Smart Phone Camera using MediaPipe and its Application in Robotics","authors":"S. Sreenath, D. Daniels, Apparaju S. D. Ganesh, Yashaswi S. Kuruganti, Rajeevlochana G. Chittawadigi","doi":"10.1109/R10-HTC53172.2021.9641542","DOIUrl":null,"url":null,"abstract":"With Industry 4.0, robots are finding their way into the production lines of multiple industries. The conventional methods to program these robots require special training, help from a robot technician, or have a basic understanding of robot kinematics. Integration of machine learning into robotics can make the programming process shorter, cost-effective and user friendly. In this paper, we propose to use a customizable machine learning solution of ‘Google MediaPipe Hands' to track human hands from a monocular camera of a smartphone and use it for robotic applications. MediaPipe Hands provides only a 2.5D pose estimation; we propose to use a simple calibration and the concept of perspective projection to get the 3D position of the hands relative to the smartphone. By conducting multiple experiments, we found that the hand tracking solution has satisfactory accuracy rates. We developed robot simulations in Python to check the viability of the hand tracker for robotic applications. We could accurately control the end-effector movement using the hand tracker.","PeriodicalId":117626,"journal":{"name":"2021 IEEE 9th Region 10 Humanitarian Technology Conference (R10-HTC)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 9th Region 10 Humanitarian Technology Conference (R10-HTC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/R10-HTC53172.2021.9641542","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
With Industry 4.0, robots are finding their way into the production lines of multiple industries. The conventional methods to program these robots require special training, help from a robot technician, or have a basic understanding of robot kinematics. Integration of machine learning into robotics can make the programming process shorter, cost-effective and user friendly. In this paper, we propose to use a customizable machine learning solution of ‘Google MediaPipe Hands' to track human hands from a monocular camera of a smartphone and use it for robotic applications. MediaPipe Hands provides only a 2.5D pose estimation; we propose to use a simple calibration and the concept of perspective projection to get the 3D position of the hands relative to the smartphone. By conducting multiple experiments, we found that the hand tracking solution has satisfactory accuracy rates. We developed robot simulations in Python to check the viability of the hand tracker for robotic applications. We could accurately control the end-effector movement using the hand tracker.