{"title":"Efficient Attention-Lightweight Deep Learning Architecture Integration for Plant Pest Recognition","authors":"Sivasubramaniam Janarthan;Selvarajah Thuseethan;Charles Joseph;Vigneshwaran Palanisamy;Sutharshan Rajasegarar;John Yearwood","doi":"10.1109/TAFE.2025.3583334","DOIUrl":null,"url":null,"abstract":"Many real-world agricultural applications, such as automatic pest recognition, benefit from lightweight deep learning (DL) architectures due to their reduced computational complexity, enabling deployment on resource-constrained devices. However, this paradigm shift comes at the cost of model performance, significantly limiting its extensive use. Traditional data-centric approaches for improving model performance, such as using large training datasets, are often unsuitable for the agricultural domain due to limited labeled data and high data collection costs. On the other hand, architectural improvements, such as attention mechanisms, have demonstrated the potential to enhance the performance of lightweight DL architectures. However, improper integration can lead to increased complexity and diminished performance. To address this challenge, this study proposes a novel mechanism to systematically determine the optimal integration configuration of popular attention techniques with the MobileNet lightweight DL architecture. The proposed method is evaluated on four variants of two benchmark plant pest datasets (D1<sub>5,869</sub> and D1<sub>500</sub>, D2<sub>1599</sub>, and D2<sub>545</sub>) and the best integration configurations are reported along with their results. The Bottleneck Attention Module (BAM) attention mechanism, integrated into 12 different layers of MobileNetV2 (BAM12), demonstrated superior performance on D1<sub>5869</sub> and D1<sub>500</sub>, and D2<sub>1599</sub> and D2<sub>545</sub>, while integrating BAM into eight layers yielded higher accuracy on D2<sub>1599</sub>. As a result, a comparison with the MobileNet baseline demonstrates that the careful integration of attention mechanisms significantly improves performance.","PeriodicalId":100637,"journal":{"name":"IEEE Transactions on AgriFood Electronics","volume":"3 2","pages":"548-560"},"PeriodicalIF":0.0000,"publicationDate":"2025-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on AgriFood Electronics","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11073813/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Many real-world agricultural applications, such as automatic pest recognition, benefit from lightweight deep learning (DL) architectures due to their reduced computational complexity, enabling deployment on resource-constrained devices. However, this paradigm shift comes at the cost of model performance, significantly limiting its extensive use. Traditional data-centric approaches for improving model performance, such as using large training datasets, are often unsuitable for the agricultural domain due to limited labeled data and high data collection costs. On the other hand, architectural improvements, such as attention mechanisms, have demonstrated the potential to enhance the performance of lightweight DL architectures. However, improper integration can lead to increased complexity and diminished performance. To address this challenge, this study proposes a novel mechanism to systematically determine the optimal integration configuration of popular attention techniques with the MobileNet lightweight DL architecture. The proposed method is evaluated on four variants of two benchmark plant pest datasets (D15,869 and D1500, D21599, and D2545) and the best integration configurations are reported along with their results. The Bottleneck Attention Module (BAM) attention mechanism, integrated into 12 different layers of MobileNetV2 (BAM12), demonstrated superior performance on D15869 and D1500, and D21599 and D2545, while integrating BAM into eight layers yielded higher accuracy on D21599. As a result, a comparison with the MobileNet baseline demonstrates that the careful integration of attention mechanisms significantly improves performance.