Zengnan Wang, Feng Yan, Liejun Wang, Yabo Yin, Jiahuan Lin
{"title":"S-YOLO:一种基于自适应门控策略和动态多尺度聚焦模块的增强小目标检测方法。","authors":"Zengnan Wang, Feng Yan, Liejun Wang, Yabo Yin, Jiahuan Lin","doi":"10.1016/j.neunet.2025.107782","DOIUrl":null,"url":null,"abstract":"<p><p>Detecting small objects in drone aerial imagery presents significant challenges, particularly when algorithms must operate in real-time under computational constraints. To address this issue, we propose S-YOLO, an efficient and streamlined small object detection framework based on YOLOv10. The S-YOLO architecture emphasizes three key innovations: (1) Enhanced Small Object Detection Layers: These layers augment semantic richness to improve detection of diminutive targets. (2) C2fGCU Module: Incorporating Gated Convolutional Units (GCU), this module adaptively modulates activation strength through deep feature analysis, enabling the model to concentrate on salient information while effectively mitigating background interference. (3) Dynamic Multi-Scale Fusion (DMSF) Module: By integrating SE-Norm with multi-scale feature extraction, this component dynamically recalibrates feature weights to optimize cross-scale information integration and focus. S-YOLO surpasses YOLOv10-n, achieving mAP50:95 improvements of 5.3%, 4.4%, and 1.4% on the VisDrone2019, AI-TOD, and DOTA1.0 datasets, respectively. Notably, S-YOLO maintains fewer parameters than YOLOv10-n while processing 285 images per second, establishing it as a highly efficient solution for real-time small object detection in aerial imagery.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"191 ","pages":"107782"},"PeriodicalIF":6.3000,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"S-YOLO: An enhanced small object detection method based on adaptive gating strategy and dynamic multi-scale focus module.\",\"authors\":\"Zengnan Wang, Feng Yan, Liejun Wang, Yabo Yin, Jiahuan Lin\",\"doi\":\"10.1016/j.neunet.2025.107782\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Detecting small objects in drone aerial imagery presents significant challenges, particularly when algorithms must operate in real-time under computational constraints. To address this issue, we propose S-YOLO, an efficient and streamlined small object detection framework based on YOLOv10. The S-YOLO architecture emphasizes three key innovations: (1) Enhanced Small Object Detection Layers: These layers augment semantic richness to improve detection of diminutive targets. (2) C2fGCU Module: Incorporating Gated Convolutional Units (GCU), this module adaptively modulates activation strength through deep feature analysis, enabling the model to concentrate on salient information while effectively mitigating background interference. (3) Dynamic Multi-Scale Fusion (DMSF) Module: By integrating SE-Norm with multi-scale feature extraction, this component dynamically recalibrates feature weights to optimize cross-scale information integration and focus. S-YOLO surpasses YOLOv10-n, achieving mAP50:95 improvements of 5.3%, 4.4%, and 1.4% on the VisDrone2019, AI-TOD, and DOTA1.0 datasets, respectively. Notably, S-YOLO maintains fewer parameters than YOLOv10-n while processing 285 images per second, establishing it as a highly efficient solution for real-time small object detection in aerial imagery.</p>\",\"PeriodicalId\":49763,\"journal\":{\"name\":\"Neural Networks\",\"volume\":\"191 \",\"pages\":\"107782\"},\"PeriodicalIF\":6.3000,\"publicationDate\":\"2025-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1016/j.neunet.2025.107782\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/7/1 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1016/j.neunet.2025.107782","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/7/1 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
S-YOLO: An enhanced small object detection method based on adaptive gating strategy and dynamic multi-scale focus module.
Detecting small objects in drone aerial imagery presents significant challenges, particularly when algorithms must operate in real-time under computational constraints. To address this issue, we propose S-YOLO, an efficient and streamlined small object detection framework based on YOLOv10. The S-YOLO architecture emphasizes three key innovations: (1) Enhanced Small Object Detection Layers: These layers augment semantic richness to improve detection of diminutive targets. (2) C2fGCU Module: Incorporating Gated Convolutional Units (GCU), this module adaptively modulates activation strength through deep feature analysis, enabling the model to concentrate on salient information while effectively mitigating background interference. (3) Dynamic Multi-Scale Fusion (DMSF) Module: By integrating SE-Norm with multi-scale feature extraction, this component dynamically recalibrates feature weights to optimize cross-scale information integration and focus. S-YOLO surpasses YOLOv10-n, achieving mAP50:95 improvements of 5.3%, 4.4%, and 1.4% on the VisDrone2019, AI-TOD, and DOTA1.0 datasets, respectively. Notably, S-YOLO maintains fewer parameters than YOLOv10-n while processing 285 images per second, establishing it as a highly efficient solution for real-time small object detection in aerial imagery.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.