Jintao Liu, Wenjin Xue, Wei Liu, Guowei Xu, Jian Liu
{"title":"Enhanced YOLOv7 for EMU Damage Detection: Overcoming False Detection and Data Scarcity by Network Optimization and AIGC","authors":"Jintao Liu, Wenjin Xue, Wei Liu, Guowei Xu, Jian Liu","doi":"10.1002/cpe.70200","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>In the context of advancements in Artificial Intelligence Generated Content (AIGC) technology, this study focuses on addressing the challenges of data scarcity and high false negative rates in the detection of damage for Electric Multiple Units (EMU). To overcome these challenges, we propose an enhanced You Only Look Once version 7 (YOLOv7) model. First, we utilize the Low-Rank Adaptation (LoRA) lightweight training strategy, which utilizes a small number of actual damage images (such as foreign objects, oil leaks, and scratches). We also use the Stable Diffusion model to generate additional high-quality damage samples, effectively enriching the training dataset. Second, we incorporate the Coordinate Concatenated Spatial(CS) Attention mechanism into the YOLOv7 backbone network, adaptively adjusting channel and spatial attention weights to improve the model's ability to detect targets in complex backgrounds while maintaining a lightweight design. Third, we decouple the detection head to independently perform classification and localization tasks. Finally, we introduce the Focal-Efficient Intersection over Union (Focal-EIOU) loss function to optimize gradient allocation during training, promoting rapid convergence of high-quality anchor boxes and improving bounding box prediction accuracy. Experiments conducted on a dataset of 3717 generated images demonstrate that the improved YOLOv7 model achieves Mean Average Precision (mAP) and Recall rates of 98.0% and 96.1%, respectively, representing significant improvements over other YOLO models.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"37 18-20","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70200","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
In the context of advancements in Artificial Intelligence Generated Content (AIGC) technology, this study focuses on addressing the challenges of data scarcity and high false negative rates in the detection of damage for Electric Multiple Units (EMU). To overcome these challenges, we propose an enhanced You Only Look Once version 7 (YOLOv7) model. First, we utilize the Low-Rank Adaptation (LoRA) lightweight training strategy, which utilizes a small number of actual damage images (such as foreign objects, oil leaks, and scratches). We also use the Stable Diffusion model to generate additional high-quality damage samples, effectively enriching the training dataset. Second, we incorporate the Coordinate Concatenated Spatial(CS) Attention mechanism into the YOLOv7 backbone network, adaptively adjusting channel and spatial attention weights to improve the model's ability to detect targets in complex backgrounds while maintaining a lightweight design. Third, we decouple the detection head to independently perform classification and localization tasks. Finally, we introduce the Focal-Efficient Intersection over Union (Focal-EIOU) loss function to optimize gradient allocation during training, promoting rapid convergence of high-quality anchor boxes and improving bounding box prediction accuracy. Experiments conducted on a dataset of 3717 generated images demonstrate that the improved YOLOv7 model achieves Mean Average Precision (mAP) and Recall rates of 98.0% and 96.1%, respectively, representing significant improvements over other YOLO models.
在人工智能生成内容(AIGC)技术进步的背景下,本研究侧重于解决电力多单元(EMU)损伤检测中数据稀缺和高假阴性率的挑战。为了克服这些挑战,我们提出了一个增强的You Only Look Once version 7 (YOLOv7)模型。首先,我们利用低秩自适应(LoRA)轻量级训练策略,该策略利用少量实际损坏图像(如异物,漏油和划痕)。我们还使用稳定扩散模型生成额外的高质量损伤样本,有效地丰富了训练数据集。其次,我们将坐标连接空间(CS)注意机制引入到YOLOv7骨干网中,自适应调整通道和空间注意权重,以提高模型在复杂背景下检测目标的能力,同时保持轻量化设计。第三,对检测头进行解耦,独立执行分类和定位任务。最后,我们引入Focal-Efficient Intersection over Union (Focal-EIOU)损失函数来优化训练过程中的梯度分配,促进高质量锚盒的快速收敛,提高边界盒的预测精度。在3717张生成图像的数据集上进行的实验表明,改进的YOLOv7模型的Mean Average Precision (mAP)和Recall(召回率)分别达到98.0%和96.1%,比其他YOLO模型有显著提高。
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.