{"title":"利用 YOLOV8 进行迁移学习,实现美国手语字母的实时识别系统","authors":"Bader Alsharif , Easa Alalwany , Mohammad Ilyas","doi":"10.1016/j.fraope.2024.100165","DOIUrl":null,"url":null,"abstract":"<div><div>Sign language serves as a sophisticated means of communication vital to individuals who are deaf or hard of hearing, relying on hand movements, facial expressions, and body language to convey nuanced meaning. American Sign Language (ASL) exemplifies this linguistic complexity with its distinct grammar and syntax. The advancement of real-time ASL gesture recognition has explored diverse methodologies, including motion sensors and computer vision techniques. This study specifically addresses the recognition of ASL alphabet gestures using computer vision through Mediapipe for hand movement tracking and YOLOv8 for training the deep learning model. The model achieved notable performance metrics: precision of 98%, recall rate of 98%, F1 score of 99%, mean Average Precision (mAP) of 98%, and mAP50-95 of 93%, underscoring its exceptional accuracy and sturdy capabilities.</div></div>","PeriodicalId":100554,"journal":{"name":"Franklin Open","volume":"8 ","pages":"Article 100165"},"PeriodicalIF":0.0000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Transfer learning with YOLOV8 for real-time recognition system of American Sign Language Alphabet\",\"authors\":\"Bader Alsharif , Easa Alalwany , Mohammad Ilyas\",\"doi\":\"10.1016/j.fraope.2024.100165\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Sign language serves as a sophisticated means of communication vital to individuals who are deaf or hard of hearing, relying on hand movements, facial expressions, and body language to convey nuanced meaning. American Sign Language (ASL) exemplifies this linguistic complexity with its distinct grammar and syntax. The advancement of real-time ASL gesture recognition has explored diverse methodologies, including motion sensors and computer vision techniques. This study specifically addresses the recognition of ASL alphabet gestures using computer vision through Mediapipe for hand movement tracking and YOLOv8 for training the deep learning model. The model achieved notable performance metrics: precision of 98%, recall rate of 98%, F1 score of 99%, mean Average Precision (mAP) of 98%, and mAP50-95 of 93%, underscoring its exceptional accuracy and sturdy capabilities.</div></div>\",\"PeriodicalId\":100554,\"journal\":{\"name\":\"Franklin Open\",\"volume\":\"8 \",\"pages\":\"Article 100165\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Franklin Open\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2773186324000951\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Franklin Open","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2773186324000951","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
手语是一种复杂的交流方式,对耳聋或听力障碍者至关重要,它依靠手部动作、面部表情和肢体语言来传达细微的含义。美国手语(ASL)以其独特的语法和句法体现了这种语言的复杂性。实时 ASL 手势识别的发展探索了多种方法,包括运动传感器和计算机视觉技术。本研究通过 Mediapipe 进行手部运动跟踪,并使用 YOLOv8 训练深度学习模型,利用计算机视觉技术专门解决 ASL 字母手势的识别问题。该模型取得了显著的性能指标:精确度 98%、召回率 98%、F1 分数 99%、平均精确度 (mAP) 98%、mAP50-95 93%,彰显了其卓越的准确性和坚固性。
Transfer learning with YOLOV8 for real-time recognition system of American Sign Language Alphabet
Sign language serves as a sophisticated means of communication vital to individuals who are deaf or hard of hearing, relying on hand movements, facial expressions, and body language to convey nuanced meaning. American Sign Language (ASL) exemplifies this linguistic complexity with its distinct grammar and syntax. The advancement of real-time ASL gesture recognition has explored diverse methodologies, including motion sensors and computer vision techniques. This study specifically addresses the recognition of ASL alphabet gestures using computer vision through Mediapipe for hand movement tracking and YOLOv8 for training the deep learning model. The model achieved notable performance metrics: precision of 98%, recall rate of 98%, F1 score of 99%, mean Average Precision (mAP) of 98%, and mAP50-95 of 93%, underscoring its exceptional accuracy and sturdy capabilities.