茶芽 DG:基于动态检测头和自适应损失函数的轻量级茶芽检测模型

IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY
Lu Jianqiang , Luo Haoxuan , Yu Chaoran , Liang Xiao , Huang Jiewei , Wu Haiwei , Wang Liang , Yang Caijuan
{"title":"茶芽 DG:基于动态检测头和自适应损失函数的轻量级茶芽检测模型","authors":"Lu Jianqiang ,&nbsp;Luo Haoxuan ,&nbsp;Yu Chaoran ,&nbsp;Liang Xiao ,&nbsp;Huang Jiewei ,&nbsp;Wu Haiwei ,&nbsp;Wang Liang ,&nbsp;Yang Caijuan","doi":"10.1016/j.compag.2024.109522","DOIUrl":null,"url":null,"abstract":"<div><div>Tea bud detection plays a crucial role in early-stage tea production estimation and robotic harvesting, significantly advancing the integration of computer vision and agriculture. Currently, tea bud detection faces several challenges such as reduced accuracy due to high background similarity, and the large size and parameter count of the models, which hinder deployment on mobile devices. To address these issues, this study introduces the lightweight Tea Bud DG model, characterized by the following features: 1) The model employs a Dynamic Head (DyHead), which enhances tea bud feature extraction through three types of perceptual attention mechanisms—scale, spatial, and task awareness. Scale awareness enables the model to adapt to objects of varying sizes; spatial awareness focuses on discriminative regions to distinguish tea buds against complex backgrounds; task awareness optimizes feature channels for specific tasks, such as classification or localization of tea buds. 2) A lightweight C3ghost module is designed, initially generating basic feature maps with fewer filters, followed by simple linear operations (e.g., translation or rotation) to create additional “ghost” feature maps, thus reducing the parameter count and model size, facilitating deployment on lightweight mobile devices. 3) By introducing the α-CIoU loss function with the parameter α, the loss and gradient of objects with different IoU scores can be adaptively reweighted by adjusting the α parameter. This approach emphasizes objects with higher IoU, enhancing the ability to identify tea buds in environments with high background similarity. The use of α-CIoU focuses on accurately differentiating tea buds from surrounding leaves, improving detection performance. The experimental results show that compared with YOLOv5s, the Tea Bud DG model reduces the model size by 31.41 % and the number of parameters by 32.21 %. Compared with YOLOv7_tiny, the size and parameters are reduced by 18.94 % and 23.84 %, respectively. It achieved improvements in [email protected] by 3 %, 3.9 %, and 5.1 %, and in [email protected]_0.95 by 2.6 %, 3.2 %, and 4 % compared with YOLOv5s, YOLOv8s, and YOLOv9s, respectively. The Tea Bud DG model estimates the tea yield with an error range of 10 % to 16 %, providing valuable data support for tea plantation management.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"227 ","pages":"Article 109522"},"PeriodicalIF":7.7000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Tea bud DG: A lightweight tea bud detection model based on dynamic detection head and adaptive loss function\",\"authors\":\"Lu Jianqiang ,&nbsp;Luo Haoxuan ,&nbsp;Yu Chaoran ,&nbsp;Liang Xiao ,&nbsp;Huang Jiewei ,&nbsp;Wu Haiwei ,&nbsp;Wang Liang ,&nbsp;Yang Caijuan\",\"doi\":\"10.1016/j.compag.2024.109522\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Tea bud detection plays a crucial role in early-stage tea production estimation and robotic harvesting, significantly advancing the integration of computer vision and agriculture. Currently, tea bud detection faces several challenges such as reduced accuracy due to high background similarity, and the large size and parameter count of the models, which hinder deployment on mobile devices. To address these issues, this study introduces the lightweight Tea Bud DG model, characterized by the following features: 1) The model employs a Dynamic Head (DyHead), which enhances tea bud feature extraction through three types of perceptual attention mechanisms—scale, spatial, and task awareness. Scale awareness enables the model to adapt to objects of varying sizes; spatial awareness focuses on discriminative regions to distinguish tea buds against complex backgrounds; task awareness optimizes feature channels for specific tasks, such as classification or localization of tea buds. 2) A lightweight C3ghost module is designed, initially generating basic feature maps with fewer filters, followed by simple linear operations (e.g., translation or rotation) to create additional “ghost” feature maps, thus reducing the parameter count and model size, facilitating deployment on lightweight mobile devices. 3) By introducing the α-CIoU loss function with the parameter α, the loss and gradient of objects with different IoU scores can be adaptively reweighted by adjusting the α parameter. This approach emphasizes objects with higher IoU, enhancing the ability to identify tea buds in environments with high background similarity. The use of α-CIoU focuses on accurately differentiating tea buds from surrounding leaves, improving detection performance. The experimental results show that compared with YOLOv5s, the Tea Bud DG model reduces the model size by 31.41 % and the number of parameters by 32.21 %. Compared with YOLOv7_tiny, the size and parameters are reduced by 18.94 % and 23.84 %, respectively. It achieved improvements in [email protected] by 3 %, 3.9 %, and 5.1 %, and in [email protected]_0.95 by 2.6 %, 3.2 %, and 4 % compared with YOLOv5s, YOLOv8s, and YOLOv9s, respectively. The Tea Bud DG model estimates the tea yield with an error range of 10 % to 16 %, providing valuable data support for tea plantation management.</div></div>\",\"PeriodicalId\":50627,\"journal\":{\"name\":\"Computers and Electronics in Agriculture\",\"volume\":\"227 \",\"pages\":\"Article 109522\"},\"PeriodicalIF\":7.7000,\"publicationDate\":\"2024-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computers and Electronics in Agriculture\",\"FirstCategoryId\":\"97\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S016816992400913X\",\"RegionNum\":1,\"RegionCategory\":\"农林科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AGRICULTURE, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers and Electronics in Agriculture","FirstCategoryId":"97","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S016816992400913X","RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

茶芽检测在早期茶叶产量估算和机器人采摘中发挥着至关重要的作用,极大地推动了计算机视觉与农业的融合。目前,茶芽检测面临着一些挑战,如背景相似度高导致精度降低,模型体积大、参数多,阻碍了在移动设备上的部署。为解决这些问题,本研究引入了轻量级茶芽 DG 模型,该模型具有以下特点:1) 该模型采用动态头(DyHead),通过三种感知注意力机制--规模意识、空间意识和任务意识,增强茶芽特征提取。尺度感知使模型能够适应不同大小的物体;空间感知集中在分辨区域,以区分复杂背景下的茶芽;任务感知优化了特定任务的特征通道,如茶芽的分类或定位。2) 设计了一个轻量级的 C3ghost 模块,该模块最初使用较少的过滤器生成基本的特征图,然后通过简单的线性操作(如平移或旋转)生成额外的 "幽灵 "特征图,从而减少了参数数量和模型大小,便于在轻量级移动设备上部署。3) 通过引入带有参数 α 的 α-CIoU 损失函数,可以通过调整 α 参数对不同 IoU 分数的对象的损失和梯度进行自适应的重新加权。这种方法强调了 IoU 值较高的对象,提高了在背景相似度较高的环境中识别茶芽的能力。α-CIoU 的使用侧重于准确区分茶芽和周围的茶叶,从而提高检测性能。实验结果表明,与 YOLOv5s 相比,茶芽 DG 模型的模型大小减少了 31.41%,参数数量减少了 32.21%。与 YOLOv7_tiny 相比,模型大小和参数数量分别减少了 18.94 % 和 23.84 %。与 YOLOv5s、YOLOv8s 和 YOLOv9s 相比,[email protected]分别提高了 3 %、3.9 % 和 5.1 %,[email protected]_0.95 分别提高了 2.6 %、3.2 % 和 4 %。茶芽 DG 模型估计的茶叶产量误差范围在 10 % 到 16 % 之间,为茶园管理提供了宝贵的数据支持。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Tea bud DG: A lightweight tea bud detection model based on dynamic detection head and adaptive loss function
Tea bud detection plays a crucial role in early-stage tea production estimation and robotic harvesting, significantly advancing the integration of computer vision and agriculture. Currently, tea bud detection faces several challenges such as reduced accuracy due to high background similarity, and the large size and parameter count of the models, which hinder deployment on mobile devices. To address these issues, this study introduces the lightweight Tea Bud DG model, characterized by the following features: 1) The model employs a Dynamic Head (DyHead), which enhances tea bud feature extraction through three types of perceptual attention mechanisms—scale, spatial, and task awareness. Scale awareness enables the model to adapt to objects of varying sizes; spatial awareness focuses on discriminative regions to distinguish tea buds against complex backgrounds; task awareness optimizes feature channels for specific tasks, such as classification or localization of tea buds. 2) A lightweight C3ghost module is designed, initially generating basic feature maps with fewer filters, followed by simple linear operations (e.g., translation or rotation) to create additional “ghost” feature maps, thus reducing the parameter count and model size, facilitating deployment on lightweight mobile devices. 3) By introducing the α-CIoU loss function with the parameter α, the loss and gradient of objects with different IoU scores can be adaptively reweighted by adjusting the α parameter. This approach emphasizes objects with higher IoU, enhancing the ability to identify tea buds in environments with high background similarity. The use of α-CIoU focuses on accurately differentiating tea buds from surrounding leaves, improving detection performance. The experimental results show that compared with YOLOv5s, the Tea Bud DG model reduces the model size by 31.41 % and the number of parameters by 32.21 %. Compared with YOLOv7_tiny, the size and parameters are reduced by 18.94 % and 23.84 %, respectively. It achieved improvements in [email protected] by 3 %, 3.9 %, and 5.1 %, and in [email protected]_0.95 by 2.6 %, 3.2 %, and 4 % compared with YOLOv5s, YOLOv8s, and YOLOv9s, respectively. The Tea Bud DG model estimates the tea yield with an error range of 10 % to 16 %, providing valuable data support for tea plantation management.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Computers and Electronics in Agriculture
Computers and Electronics in Agriculture 工程技术-计算机:跨学科应用
CiteScore
15.30
自引率
14.50%
发文量
800
审稿时长
62 days
期刊介绍: Computers and Electronics in Agriculture provides international coverage of advancements in computer hardware, software, electronic instrumentation, and control systems applied to agricultural challenges. Encompassing agronomy, horticulture, forestry, aquaculture, and animal farming, the journal publishes original papers, reviews, and applications notes. It explores the use of computers and electronics in plant or animal agricultural production, covering topics like agricultural soils, water, pests, controlled environments, and waste. The scope extends to on-farm post-harvest operations and relevant technologies, including artificial intelligence, sensors, machine vision, robotics, networking, and simulation modeling. Its companion journal, Smart Agricultural Technology, continues the focus on smart applications in production agriculture.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信