{"title":"Recognition of Forest Fire Smoke Based on Improved YOLOv8n Model","authors":"Faying Chen, Meng Yang, Yuan Wang","doi":"10.1007/s10694-025-01733-x","DOIUrl":null,"url":null,"abstract":"<div><p>To address the challenges of early forest fire smoke image recognition, including false alarms and missed reports caused by interference in complex environments, an enhanced model, named MB-YOLO, is proposed based on the YOLOv8 Nano (YOLOv8n) architecture for efficiently recognizing forest fire smoke. Firstly, to overcome detection failures of low-concentration smoke in complex backgrounds, the original Path Aggregation Network (PAN) is replaced with a bi-directional feature pyramid network (BiFPN). This substitution not only enhances multi-scale feature extraction but also simplifies the network structure, reducing the number of parameters. Secondly, to address false detections caused by cloud and mist interference, the C2f_MLCA module is developed. This module integrates a lightweight Mixed Local Attention mechanism (MLCA) into the bottleneck of the gradient flow module C2f, thereby enhancing smoke feature extraction. Lastly, to reduce sensitivity to positional offsets of small smoke targets, the Complete Intersection over Union (CIoU) loss is replaced with Inner-DIoU loss. This new loss function computes loss with auxiliary bounding boxes, accelerating convergence speed and enhancing accuracy for small smoke targets. The effectiveness of the algorithm is validated with a curated dataset containing small smoke targets, unclear backlighting, and cloud and mist interference. Experimental results demonstrate that our model achieves a mean Average Precision (mAP) of 80.1%, a frame rate of 60.6 Frames Per Second (FPS), with a total of 1.09 million parameters and 7.1 billion floating-point operations per second (FLOPs). This model offers high detection accuracy, fewer parameters, and lower GFLOPs, facilitating accurate real-time monitoring of forest fires in complex environments and all-weather conditions.</p></div>","PeriodicalId":558,"journal":{"name":"Fire Technology","volume":"61 5","pages":"3351 - 3374"},"PeriodicalIF":2.4000,"publicationDate":"2025-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Fire Technology","FirstCategoryId":"5","ListUrlMain":"https://link.springer.com/article/10.1007/s10694-025-01733-x","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
To address the challenges of early forest fire smoke image recognition, including false alarms and missed reports caused by interference in complex environments, an enhanced model, named MB-YOLO, is proposed based on the YOLOv8 Nano (YOLOv8n) architecture for efficiently recognizing forest fire smoke. Firstly, to overcome detection failures of low-concentration smoke in complex backgrounds, the original Path Aggregation Network (PAN) is replaced with a bi-directional feature pyramid network (BiFPN). This substitution not only enhances multi-scale feature extraction but also simplifies the network structure, reducing the number of parameters. Secondly, to address false detections caused by cloud and mist interference, the C2f_MLCA module is developed. This module integrates a lightweight Mixed Local Attention mechanism (MLCA) into the bottleneck of the gradient flow module C2f, thereby enhancing smoke feature extraction. Lastly, to reduce sensitivity to positional offsets of small smoke targets, the Complete Intersection over Union (CIoU) loss is replaced with Inner-DIoU loss. This new loss function computes loss with auxiliary bounding boxes, accelerating convergence speed and enhancing accuracy for small smoke targets. The effectiveness of the algorithm is validated with a curated dataset containing small smoke targets, unclear backlighting, and cloud and mist interference. Experimental results demonstrate that our model achieves a mean Average Precision (mAP) of 80.1%, a frame rate of 60.6 Frames Per Second (FPS), with a total of 1.09 million parameters and 7.1 billion floating-point operations per second (FLOPs). This model offers high detection accuracy, fewer parameters, and lower GFLOPs, facilitating accurate real-time monitoring of forest fires in complex environments and all-weather conditions.
期刊介绍:
Fire Technology publishes original contributions, both theoretical and empirical, that contribute to the solution of problems in fire safety science and engineering. It is the leading journal in the field, publishing applied research dealing with the full range of actual and potential fire hazards facing humans and the environment. It covers the entire domain of fire safety science and engineering problems relevant in industrial, operational, cultural, and environmental applications, including modeling, testing, detection, suppression, human behavior, wildfires, structures, and risk analysis.
The aim of Fire Technology is to push forward the frontiers of knowledge and technology by encouraging interdisciplinary communication of significant technical developments in fire protection and subjects of scientific interest to the fire protection community at large.
It is published in conjunction with the National Fire Protection Association (NFPA) and the Society of Fire Protection Engineers (SFPE). The mission of NFPA is to help save lives and reduce loss with information, knowledge, and passion. The mission of SFPE is advancing the science and practice of fire protection engineering internationally.