May Regev, Avital Bechar, Yuval Cohen, Avraham Sadowsky, Sigal Berman
{"title":"Instance segmentation of partially occluded Medjool-date fruit bunches for robotic thinning","authors":"May Regev, Avital Bechar, Yuval Cohen, Avraham Sadowsky, Sigal Berman","doi":"10.1007/s11119-023-10086-w","DOIUrl":null,"url":null,"abstract":"<p>Medjool date thinning automation is essential for reducing Medjool production labor and improving fruit quality. Thinning automation requires motion planning based on feature extraction from a segmented fruit bunch and its components. Previous research with focused bunch images attained high success in bunch component segmentation but less success in establishing correct association between the two components (a rachis and spikelets) that form one bunch. The current study presents an algorithm for improved component segmentation and association in the presence of occlusions based on integrating deep neural networks, traditional methods building on bunch geometry, and active vision. Following segmentation with Mask-R-CNN, segmented component images are converted to binary images with a Savitzky–Golay filter and an adapted Otsu threshold. Bunch orientation is calculated based on lines found in the binary image with the Hough transform. The orientation is used for associating a rachis with spikelets. If a suitable rachis is not found, bunch orientation is used for selecting a better viewpoint. The method was tested with two databases of bunches in an orchard, one with focused and one with non-focused images. In all images, the spikelets were correctly identified [intersection over union (IoU) 0.5: F1 0.9]. The average orientation errors were 18.15° (SD 12.77°) and 16.44° (SD 11.07°), respectively, for the focused and non-focused databases. For correct rachis selection, precision was very high when incorporating orientation, and when additionally incorporating active vision recall (and therefore F1) was high (IoU 0.5: orientation: precision 0.94, recall 0.44, F1 0.60; addition of active vision: precision 0.96, recall 0.61, F1 0.74). The developed method leads to highly accurate identification of fruit bunches and their spikelets and rachis, making it suitable for integration with a thinning automation system.</p>","PeriodicalId":20423,"journal":{"name":"Precision Agriculture","volume":"3 1","pages":""},"PeriodicalIF":5.4000,"publicationDate":"2023-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Precision Agriculture","FirstCategoryId":"97","ListUrlMain":"https://doi.org/10.1007/s11119-023-10086-w","RegionNum":2,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
Medjool date thinning automation is essential for reducing Medjool production labor and improving fruit quality. Thinning automation requires motion planning based on feature extraction from a segmented fruit bunch and its components. Previous research with focused bunch images attained high success in bunch component segmentation but less success in establishing correct association between the two components (a rachis and spikelets) that form one bunch. The current study presents an algorithm for improved component segmentation and association in the presence of occlusions based on integrating deep neural networks, traditional methods building on bunch geometry, and active vision. Following segmentation with Mask-R-CNN, segmented component images are converted to binary images with a Savitzky–Golay filter and an adapted Otsu threshold. Bunch orientation is calculated based on lines found in the binary image with the Hough transform. The orientation is used for associating a rachis with spikelets. If a suitable rachis is not found, bunch orientation is used for selecting a better viewpoint. The method was tested with two databases of bunches in an orchard, one with focused and one with non-focused images. In all images, the spikelets were correctly identified [intersection over union (IoU) 0.5: F1 0.9]. The average orientation errors were 18.15° (SD 12.77°) and 16.44° (SD 11.07°), respectively, for the focused and non-focused databases. For correct rachis selection, precision was very high when incorporating orientation, and when additionally incorporating active vision recall (and therefore F1) was high (IoU 0.5: orientation: precision 0.94, recall 0.44, F1 0.60; addition of active vision: precision 0.96, recall 0.61, F1 0.74). The developed method leads to highly accurate identification of fruit bunches and their spikelets and rachis, making it suitable for integration with a thinning automation system.
期刊介绍:
Precision Agriculture promotes the most innovative results coming from the research in the field of precision agriculture. It provides an effective forum for disseminating original and fundamental research and experience in the rapidly advancing area of precision farming.
There are many topics in the field of precision agriculture; therefore, the topics that are addressed include, but are not limited to:
Natural Resources Variability: Soil and landscape variability, digital elevation models, soil mapping, geostatistics, geographic information systems, microclimate, weather forecasting, remote sensing, management units, scale, etc.
Managing Variability: Sampling techniques, site-specific nutrient and crop protection chemical recommendation, crop quality, tillage, seed density, seed variety, yield mapping, remote sensing, record keeping systems, data interpretation and use, crops (corn, wheat, sugar beets, potatoes, peanut, cotton, vegetables, etc.), management scale, etc.
Engineering Technology: Computers, positioning systems, DGPS, machinery, tillage, planting, nutrient and crop protection implements, manure, irrigation, fertigation, yield monitor and mapping, soil physical and chemical characteristic sensors, weed/pest mapping, etc.
Profitability: MEY, net returns, BMPs, optimum recommendations, crop quality, technology cost, sustainability, social impacts, marketing, cooperatives, farm scale, crop type, etc.
Environment: Nutrient, crop protection chemicals, sediments, leaching, runoff, practices, field, watershed, on/off farm, artificial drainage, ground water, surface water, etc.
Technology Transfer: Skill needs, education, training, outreach, methods, surveys, agri-business, producers, distance education, Internet, simulations models, decision support systems, expert systems, on-farm experimentation, partnerships, quality of rural life, etc.