基于掩模R-CNN的第一人称图像手部检测与分割

Huy Nguyen, Hai Vu
{"title":"基于掩模R-CNN的第一人称图像手部检测与分割","authors":"Huy Nguyen, Hai Vu","doi":"10.32913/mic-ict-research.v2022.n1.1001","DOIUrl":null,"url":null,"abstract":"In this work, we propose a technique to automatically detect and segment hands on first-person images of patientsin upper limb rehabilitation exercises. The aim is to automate the assessment of the patient's recovery processthrough rehabilitation exercises. The proposed technique includes the following steps: 1) setting up a wearablecamera system and collecting upper extremity rehabilitation exercise data. The data is filtered, selected andannotated with the left and right hand as well as segmented the image area of the patient's hand. The datasetconsists of 3700 images with the name RehabHand. This dataset is used to train hand detection and segmentationmodels on first-person images. 2) conducted a survey of automatic hand detection and segmentation models usingMask-RCNN network architecture with different backbones. From the experimental architectures, the Mask -RCNN architecture with the Res2Net backbone was selected for all three tasks: hand detection; left - right handidentification; and hand segmentation. The proposed model has achieved the highest performance in the tests. Toovercome the limitation on the amount of training data, we propose to use the transfer learning method alongwith data enhancement techniques to improve the accuracy of the model. The results of the detection of objects onthe test dataset for the left hand is AP = 92.3%, the right hand AP = 91.1%. The segmentation result on the test dataset forleft hand is AP = 88.8%, right hand being AP = 87%. These results suggest that it is possible to automatically quantifythe patient's ability to use their hands during upper extremity rehabilitation.","PeriodicalId":432355,"journal":{"name":"Research and Development on Information and Communication Technology","volume":"53 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Hand Detection and Segmentation in First Person Image Using Mask R-CNN\",\"authors\":\"Huy Nguyen, Hai Vu\",\"doi\":\"10.32913/mic-ict-research.v2022.n1.1001\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this work, we propose a technique to automatically detect and segment hands on first-person images of patientsin upper limb rehabilitation exercises. The aim is to automate the assessment of the patient's recovery processthrough rehabilitation exercises. The proposed technique includes the following steps: 1) setting up a wearablecamera system and collecting upper extremity rehabilitation exercise data. The data is filtered, selected andannotated with the left and right hand as well as segmented the image area of the patient's hand. The datasetconsists of 3700 images with the name RehabHand. This dataset is used to train hand detection and segmentationmodels on first-person images. 2) conducted a survey of automatic hand detection and segmentation models usingMask-RCNN network architecture with different backbones. From the experimental architectures, the Mask -RCNN architecture with the Res2Net backbone was selected for all three tasks: hand detection; left - right handidentification; and hand segmentation. The proposed model has achieved the highest performance in the tests. Toovercome the limitation on the amount of training data, we propose to use the transfer learning method alongwith data enhancement techniques to improve the accuracy of the model. The results of the detection of objects onthe test dataset for the left hand is AP = 92.3%, the right hand AP = 91.1%. The segmentation result on the test dataset forleft hand is AP = 88.8%, right hand being AP = 87%. These results suggest that it is possible to automatically quantifythe patient's ability to use their hands during upper extremity rehabilitation.\",\"PeriodicalId\":432355,\"journal\":{\"name\":\"Research and Development on Information and Communication Technology\",\"volume\":\"53 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-03-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Research and Development on Information and Communication Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.32913/mic-ict-research.v2022.n1.1001\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research and Development on Information and Communication Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.32913/mic-ict-research.v2022.n1.1001","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

在这项工作中,我们提出了一种自动检测和分割上肢康复训练患者第一人称图像上的手的技术。其目的是通过康复训练来自动评估病人的康复过程。提出的技术包括以下步骤:1)建立可穿戴相机系统,收集上肢康复运动数据。用左手和右手对数据进行过滤、选择和标注,并对患者手部的图像区域进行分割。该数据集由3700张名为RehabHand的图像组成。该数据集用于训练第一人称图像上的手部检测和分割模型。2)对基于不同主干的mask - rcnn网络架构的手部自动检测与分割模型进行了研究。从实验架构来看,我们选择了具有Res2Net主干的Mask -RCNN架构来完成所有三个任务:手部检测;左-右手识别;手分割。该模型在测试中取得了最高的性能。为了克服训练数据量的限制,我们提出使用迁移学习方法和数据增强技术来提高模型的准确性。左手在测试数据集中检测物体的结果为AP = 92.3%,右手AP = 91.1%。在测试数据集上,左手的分割结果为AP = 88.8%,右手为AP = 87%。这些结果表明,在上肢康复过程中,可以自动量化患者使用双手的能力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Hand Detection and Segmentation in First Person Image Using Mask R-CNN
In this work, we propose a technique to automatically detect and segment hands on first-person images of patientsin upper limb rehabilitation exercises. The aim is to automate the assessment of the patient's recovery processthrough rehabilitation exercises. The proposed technique includes the following steps: 1) setting up a wearablecamera system and collecting upper extremity rehabilitation exercise data. The data is filtered, selected andannotated with the left and right hand as well as segmented the image area of the patient's hand. The datasetconsists of 3700 images with the name RehabHand. This dataset is used to train hand detection and segmentationmodels on first-person images. 2) conducted a survey of automatic hand detection and segmentation models usingMask-RCNN network architecture with different backbones. From the experimental architectures, the Mask -RCNN architecture with the Res2Net backbone was selected for all three tasks: hand detection; left - right handidentification; and hand segmentation. The proposed model has achieved the highest performance in the tests. Toovercome the limitation on the amount of training data, we propose to use the transfer learning method alongwith data enhancement techniques to improve the accuracy of the model. The results of the detection of objects onthe test dataset for the left hand is AP = 92.3%, the right hand AP = 91.1%. The segmentation result on the test dataset forleft hand is AP = 88.8%, right hand being AP = 87%. These results suggest that it is possible to automatically quantifythe patient's ability to use their hands during upper extremity rehabilitation.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信