{"title":"基于 YOLOv7 的两阶段路标检测和文本识别系统","authors":"Chen-Chiung Hsieh , Chia-Hao Hsu , Wei-Hsin Huang","doi":"10.1016/j.iot.2024.101330","DOIUrl":null,"url":null,"abstract":"<div><p>We developed a two-stage traffic sign recognition system to enhance safety and prevent tragic traffic incidents involving self-driving cars. In the first stage, YOLOv7 was employed as the detection model for identifying 31 types of traffic signs. Input images were set to 640 × 640 pixels to balance speed and accuracy, with high-definition images split into overlapping sub-images of the same size for training. The YOLOv7 model achieved a training accuracy of 99.2 % and demonstrated robustness across various scenes, earning a testing accuracy of 99 % in both YouTube and self-recorded driving videos. In the second stage, extracted road sign images underwent rectification before processing with OCR tools such as EasyOCR and PaddleOCR. Post-processing steps addressed potential confusion, particularly with city/town names. After extensive testing, the system achieved recognition rates of 97.5 % for alphabets and 99.4 % for Chinese characters. This system significantly enhances the ability of self-driving cars to detect and interpret traffic signs, thereby contributing to safer road travel.</p></div>","PeriodicalId":29968,"journal":{"name":"Internet of Things","volume":null,"pages":null},"PeriodicalIF":6.0000,"publicationDate":"2024-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A two-stage road sign detection and text recognition system based on YOLOv7\",\"authors\":\"Chen-Chiung Hsieh , Chia-Hao Hsu , Wei-Hsin Huang\",\"doi\":\"10.1016/j.iot.2024.101330\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>We developed a two-stage traffic sign recognition system to enhance safety and prevent tragic traffic incidents involving self-driving cars. In the first stage, YOLOv7 was employed as the detection model for identifying 31 types of traffic signs. Input images were set to 640 × 640 pixels to balance speed and accuracy, with high-definition images split into overlapping sub-images of the same size for training. The YOLOv7 model achieved a training accuracy of 99.2 % and demonstrated robustness across various scenes, earning a testing accuracy of 99 % in both YouTube and self-recorded driving videos. In the second stage, extracted road sign images underwent rectification before processing with OCR tools such as EasyOCR and PaddleOCR. Post-processing steps addressed potential confusion, particularly with city/town names. After extensive testing, the system achieved recognition rates of 97.5 % for alphabets and 99.4 % for Chinese characters. This system significantly enhances the ability of self-driving cars to detect and interpret traffic signs, thereby contributing to safer road travel.</p></div>\",\"PeriodicalId\":29968,\"journal\":{\"name\":\"Internet of Things\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2024-08-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Internet of Things\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2542660524002713\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Internet of Things","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2542660524002713","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
A two-stage road sign detection and text recognition system based on YOLOv7
We developed a two-stage traffic sign recognition system to enhance safety and prevent tragic traffic incidents involving self-driving cars. In the first stage, YOLOv7 was employed as the detection model for identifying 31 types of traffic signs. Input images were set to 640 × 640 pixels to balance speed and accuracy, with high-definition images split into overlapping sub-images of the same size for training. The YOLOv7 model achieved a training accuracy of 99.2 % and demonstrated robustness across various scenes, earning a testing accuracy of 99 % in both YouTube and self-recorded driving videos. In the second stage, extracted road sign images underwent rectification before processing with OCR tools such as EasyOCR and PaddleOCR. Post-processing steps addressed potential confusion, particularly with city/town names. After extensive testing, the system achieved recognition rates of 97.5 % for alphabets and 99.4 % for Chinese characters. This system significantly enhances the ability of self-driving cars to detect and interpret traffic signs, thereby contributing to safer road travel.
期刊介绍:
Internet of Things; Engineering Cyber Physical Human Systems is a comprehensive journal encouraging cross collaboration between researchers, engineers and practitioners in the field of IoT & Cyber Physical Human Systems. The journal offers a unique platform to exchange scientific information on the entire breadth of technology, science, and societal applications of the IoT.
The journal will place a high priority on timely publication, and provide a home for high quality.
Furthermore, IOT is interested in publishing topical Special Issues on any aspect of IOT.