Recognition of mango and location of picking point on stem based on a multi-task CNN model named YOLOMS

IF 5.4 2区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY
Bin Zhang, Yuyang Xia, Rongrong Wang, Yong Wang, Chenghai Yin, Meng Fu, Wei Fu
{"title":"Recognition of mango and location of picking point on stem based on a multi-task CNN model named YOLOMS","authors":"Bin Zhang, Yuyang Xia, Rongrong Wang, Yong Wang, Chenghai Yin, Meng Fu, Wei Fu","doi":"10.1007/s11119-024-10119-y","DOIUrl":null,"url":null,"abstract":"<p>Due to the fact that the color of mango peel is similar to that of leaf, and there are many fruits on one stem, it is difficult to locate the picking point when using robots to pick fresh mango in the natural environment. A multi-task learning method named YOLOMS was proposed for mango recognition and rapid location of main stem picking points. Firstly, the backbone network of YOLOv5s was optimized and improved by using the RepVGG structure. The loss function of original YOLOv5s was improved by introducing the loss function of Focal-EIoU. The improved model could accurately identify mango and fruit stem in complex environment without decreasing reasoning speed. Secondly, the subtask of mango stem segmentation was added to the improved YOLOv5s model, and the YOLOMS multi-task model was constructed to obtain the location and semantic information of the fruit stem. Finally, the strategies of main fruit stem recognition and picking point location were put forward to realize the picking point location of the whole cluster mango. The images of mangoes on trees in natural environment were collected to test the performance of the YOLOMS model. The test results showed that the mAP and Recall of mango fruit and stem target detection by YOLOMS model were 82.42% and 85.64%, respectively, and the MIoU of stem semantic segmentation reached to 82.26%. The recognition accuracy of mangoes was 92.19%, the success rate of stem picking location was 89.84%, and the average location time was 58.4 ms. Compared with the target detection models of Yolov4, Yolov5s, Yolov7-tiny and the target segmentation models of U-net, PSPNet and DeepLab_v3+, the improved YOLOMS model had significantly better performance, which could quickly and accurately locate the picking point. This research provides technical support for mango picking robot to recognize the fruit and locate the picking point.</p>","PeriodicalId":20423,"journal":{"name":"Precision Agriculture","volume":"305 1 1","pages":""},"PeriodicalIF":5.4000,"publicationDate":"2024-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Precision Agriculture","FirstCategoryId":"97","ListUrlMain":"https://doi.org/10.1007/s11119-024-10119-y","RegionNum":2,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Due to the fact that the color of mango peel is similar to that of leaf, and there are many fruits on one stem, it is difficult to locate the picking point when using robots to pick fresh mango in the natural environment. A multi-task learning method named YOLOMS was proposed for mango recognition and rapid location of main stem picking points. Firstly, the backbone network of YOLOv5s was optimized and improved by using the RepVGG structure. The loss function of original YOLOv5s was improved by introducing the loss function of Focal-EIoU. The improved model could accurately identify mango and fruit stem in complex environment without decreasing reasoning speed. Secondly, the subtask of mango stem segmentation was added to the improved YOLOv5s model, and the YOLOMS multi-task model was constructed to obtain the location and semantic information of the fruit stem. Finally, the strategies of main fruit stem recognition and picking point location were put forward to realize the picking point location of the whole cluster mango. The images of mangoes on trees in natural environment were collected to test the performance of the YOLOMS model. The test results showed that the mAP and Recall of mango fruit and stem target detection by YOLOMS model were 82.42% and 85.64%, respectively, and the MIoU of stem semantic segmentation reached to 82.26%. The recognition accuracy of mangoes was 92.19%, the success rate of stem picking location was 89.84%, and the average location time was 58.4 ms. Compared with the target detection models of Yolov4, Yolov5s, Yolov7-tiny and the target segmentation models of U-net, PSPNet and DeepLab_v3+, the improved YOLOMS model had significantly better performance, which could quickly and accurately locate the picking point. This research provides technical support for mango picking robot to recognize the fruit and locate the picking point.

Abstract Image

基于名为 YOLOMS 的多任务 CNN 模型识别芒果并确定茎干上的采摘点位置
由于芒果果皮的颜色与叶片相似,且一根茎上有许多果实,因此在自然环境中使用机器人采摘新鲜芒果时,很难定位采摘点。为实现芒果识别和主茎采摘点的快速定位,提出了一种名为 YOLOMS 的多任务学习方法。首先,利用 RepVGG 结构对 YOLOv5s 的骨干网络进行了优化和改进。通过引入 Focal-EIoU 的损失函数,改进了原始 YOLOv5s 的损失函数。改进后的模型可以在不降低推理速度的情况下准确识别复杂环境中的芒果和果柄。其次,在改进的 YOLOv5s 模型中加入了芒果果柄分割子任务,并构建了 YOLOMS 多任务模型,以获取果柄的位置和语义信息。最后,提出了主果梗识别和采摘点定位策略,实现了整簇芒果的采摘点定位。为了测试 YOLOMS 模型的性能,我们收集了自然环境中树上芒果的图像。测试结果表明,YOLOMS 模型检测芒果果实和茎干目标的 mAP 和 Recall 分别为 82.42% 和 85.64%,茎干语义分割的 MIoU 达到 82.26%。芒果的识别准确率为 92.19%,茎干采摘定位的成功率为 89.84%,平均定位时间为 58.4 毫秒。与 Yolov4、Yolov5s、Yolov7-tiny 等目标检测模型和 U-net、PSPNet、DeepLab_v3+ 等目标分割模型相比,改进后的 YOLOMS 模型性能明显提高,能快速准确地定位采摘点。这项研究为芒果采摘机器人识别水果和定位采摘点提供了技术支持。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Precision Agriculture
Precision Agriculture 农林科学-农业综合
CiteScore
12.30
自引率
8.10%
发文量
103
审稿时长
>24 weeks
期刊介绍: Precision Agriculture promotes the most innovative results coming from the research in the field of precision agriculture. It provides an effective forum for disseminating original and fundamental research and experience in the rapidly advancing area of precision farming. There are many topics in the field of precision agriculture; therefore, the topics that are addressed include, but are not limited to: Natural Resources Variability: Soil and landscape variability, digital elevation models, soil mapping, geostatistics, geographic information systems, microclimate, weather forecasting, remote sensing, management units, scale, etc. Managing Variability: Sampling techniques, site-specific nutrient and crop protection chemical recommendation, crop quality, tillage, seed density, seed variety, yield mapping, remote sensing, record keeping systems, data interpretation and use, crops (corn, wheat, sugar beets, potatoes, peanut, cotton, vegetables, etc.), management scale, etc. Engineering Technology: Computers, positioning systems, DGPS, machinery, tillage, planting, nutrient and crop protection implements, manure, irrigation, fertigation, yield monitor and mapping, soil physical and chemical characteristic sensors, weed/pest mapping, etc. Profitability: MEY, net returns, BMPs, optimum recommendations, crop quality, technology cost, sustainability, social impacts, marketing, cooperatives, farm scale, crop type, etc. Environment: Nutrient, crop protection chemicals, sediments, leaching, runoff, practices, field, watershed, on/off farm, artificial drainage, ground water, surface water, etc. Technology Transfer: Skill needs, education, training, outreach, methods, surveys, agri-business, producers, distance education, Internet, simulations models, decision support systems, expert systems, on-farm experimentation, partnerships, quality of rural life, etc.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信