{"title":"CO-YOLO:一种轻量级、高效的油茶果实目标检测与姿态确定模型","authors":"Shouxiang Jin , Lei Zhou , Hongping Zhou","doi":"10.1016/j.compag.2025.110394","DOIUrl":null,"url":null,"abstract":"<div><div>The complex growth patterns of <em>Camellia oleifera</em> fruits in natural environments pose significant challenges for harvesting robots. Conventional solutions often rely on complex 3D point cloud processing to detect these growth patterns and enable robotic harvesting. In this study, a method for recognizing the growth patterns of <em>Camellia oleifera</em> fruits in 2D images is proposed. To address the challenges of precise posture detection, the fruits are categorized into five types based on their growth patterns: Front, Up, Down, Left, and Right. Occluded fruits are classified separately, and a dedicated dataset for posture recognition is created. Furthermore, a posture detection model, CO-YOLO, based on the YOLO11n architecture, is introduced. The Multi-scale Aggregation Attention (MMA) module replaces the original C3f2 module, enabling the fusion of feature information across multiple scales, which enhances the model’s perceptual capabilities and improves posture recognition accuracy. Additionally, the Depth Pointwise Convolutional (DPW) module is introduced to replace standard convolutions in the backbone and neck networks, enabling better fusion of channel features, enhancing the representation of posture features in the target region, and reducing the number of parameters. Experimental results show that CO-YOLO achieves a precision of 90.6 %, a recall of 87.0 %, and an [email protected] of 93.7 %. Compared to YOLO11s, CO-YOLO reduces model size and computational complexity by 77.1 % and 69.9 %, respectively, while improving [email protected] by 4.8 %. Compared to YOLOv7-tiny, YOLOv9s, and YOLOv10s, CO-YOLO achieves increases in [email protected] of 3.9 %, 5.6 %, and 5.5 %, respectively. Heatmaps generated by CO-YOLO and YOLO11n indicate that these enhancements significantly improve posture recognition. In summary, the CO-YOLO model exhibits strong performance and provides valuable insights for advancing fruit-picking robotics.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"235 ","pages":"Article 110394"},"PeriodicalIF":8.9000,"publicationDate":"2025-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CO-YOLO: A lightweight and efficient model for Camellia oleifera fruit object detection and posture determination\",\"authors\":\"Shouxiang Jin , Lei Zhou , Hongping Zhou\",\"doi\":\"10.1016/j.compag.2025.110394\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The complex growth patterns of <em>Camellia oleifera</em> fruits in natural environments pose significant challenges for harvesting robots. Conventional solutions often rely on complex 3D point cloud processing to detect these growth patterns and enable robotic harvesting. In this study, a method for recognizing the growth patterns of <em>Camellia oleifera</em> fruits in 2D images is proposed. To address the challenges of precise posture detection, the fruits are categorized into five types based on their growth patterns: Front, Up, Down, Left, and Right. Occluded fruits are classified separately, and a dedicated dataset for posture recognition is created. Furthermore, a posture detection model, CO-YOLO, based on the YOLO11n architecture, is introduced. The Multi-scale Aggregation Attention (MMA) module replaces the original C3f2 module, enabling the fusion of feature information across multiple scales, which enhances the model’s perceptual capabilities and improves posture recognition accuracy. Additionally, the Depth Pointwise Convolutional (DPW) module is introduced to replace standard convolutions in the backbone and neck networks, enabling better fusion of channel features, enhancing the representation of posture features in the target region, and reducing the number of parameters. Experimental results show that CO-YOLO achieves a precision of 90.6 %, a recall of 87.0 %, and an [email protected] of 93.7 %. Compared to YOLO11s, CO-YOLO reduces model size and computational complexity by 77.1 % and 69.9 %, respectively, while improving [email protected] by 4.8 %. Compared to YOLOv7-tiny, YOLOv9s, and YOLOv10s, CO-YOLO achieves increases in [email protected] of 3.9 %, 5.6 %, and 5.5 %, respectively. Heatmaps generated by CO-YOLO and YOLO11n indicate that these enhancements significantly improve posture recognition. In summary, the CO-YOLO model exhibits strong performance and provides valuable insights for advancing fruit-picking robotics.</div></div>\",\"PeriodicalId\":50627,\"journal\":{\"name\":\"Computers and Electronics in Agriculture\",\"volume\":\"235 \",\"pages\":\"Article 110394\"},\"PeriodicalIF\":8.9000,\"publicationDate\":\"2025-04-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computers and Electronics in Agriculture\",\"FirstCategoryId\":\"97\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0168169925005009\",\"RegionNum\":1,\"RegionCategory\":\"农林科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AGRICULTURE, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers and Electronics in Agriculture","FirstCategoryId":"97","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0168169925005009","RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURE, MULTIDISCIPLINARY","Score":null,"Total":0}
CO-YOLO: A lightweight and efficient model for Camellia oleifera fruit object detection and posture determination
The complex growth patterns of Camellia oleifera fruits in natural environments pose significant challenges for harvesting robots. Conventional solutions often rely on complex 3D point cloud processing to detect these growth patterns and enable robotic harvesting. In this study, a method for recognizing the growth patterns of Camellia oleifera fruits in 2D images is proposed. To address the challenges of precise posture detection, the fruits are categorized into five types based on their growth patterns: Front, Up, Down, Left, and Right. Occluded fruits are classified separately, and a dedicated dataset for posture recognition is created. Furthermore, a posture detection model, CO-YOLO, based on the YOLO11n architecture, is introduced. The Multi-scale Aggregation Attention (MMA) module replaces the original C3f2 module, enabling the fusion of feature information across multiple scales, which enhances the model’s perceptual capabilities and improves posture recognition accuracy. Additionally, the Depth Pointwise Convolutional (DPW) module is introduced to replace standard convolutions in the backbone and neck networks, enabling better fusion of channel features, enhancing the representation of posture features in the target region, and reducing the number of parameters. Experimental results show that CO-YOLO achieves a precision of 90.6 %, a recall of 87.0 %, and an [email protected] of 93.7 %. Compared to YOLO11s, CO-YOLO reduces model size and computational complexity by 77.1 % and 69.9 %, respectively, while improving [email protected] by 4.8 %. Compared to YOLOv7-tiny, YOLOv9s, and YOLOv10s, CO-YOLO achieves increases in [email protected] of 3.9 %, 5.6 %, and 5.5 %, respectively. Heatmaps generated by CO-YOLO and YOLO11n indicate that these enhancements significantly improve posture recognition. In summary, the CO-YOLO model exhibits strong performance and provides valuable insights for advancing fruit-picking robotics.
期刊介绍:
Computers and Electronics in Agriculture provides international coverage of advancements in computer hardware, software, electronic instrumentation, and control systems applied to agricultural challenges. Encompassing agronomy, horticulture, forestry, aquaculture, and animal farming, the journal publishes original papers, reviews, and applications notes. It explores the use of computers and electronics in plant or animal agricultural production, covering topics like agricultural soils, water, pests, controlled environments, and waste. The scope extends to on-farm post-harvest operations and relevant technologies, including artificial intelligence, sensors, machine vision, robotics, networking, and simulation modeling. Its companion journal, Smart Agricultural Technology, continues the focus on smart applications in production agriculture.