CNN-based damage classification of soybean kernels using a high-magnification image dataset

IF 2.9 3区 农林科学 Q2 FOOD SCIENCE & TECHNOLOGY
Isparsh Chauhan, Siddharth Kekre, Ankur Miglani, Pavan Kumar Kankar, Milind B. Ratnaparkhe
{"title":"CNN-based damage classification of soybean kernels using a high-magnification image dataset","authors":"Isparsh Chauhan,&nbsp;Siddharth Kekre,&nbsp;Ankur Miglani,&nbsp;Pavan Kumar Kankar,&nbsp;Milind B. Ratnaparkhe","doi":"10.1007/s11694-025-03195-9","DOIUrl":null,"url":null,"abstract":"<div><p>The assessment of the surface quality of pre-processed soybean kernels is crucial in determining their market acceptance, storage stability, processing quality, and overall consumer approval. Conventional techniques of surface quality evaluation are time-consuming, reliant on personal judgement, and lack consistency. Conversely, the existing techniques are restricted to either selecting healthy soybean kernels from damaged ones without categorizing the damaged ones, or separating different varieties. The lack of a labelled, high-magnification image dataset and the use of advanced CNN models have hindered the exploration of a detailed classification of damage in soybean kernels. These models excel at end-to-end tasks, minimize pre-processing, and eliminate the need for manual feature extraction, enabling quick, accurate, and precise classification. This study demonstrates the use of a machine vision system to create an image dataset consisting of 9866 high-magnification (2.85 µm/pixel) images of soybean kernels with damages. The dataset encompasses eight distinct damage classes: healthy, heat damage (HD), immature damage (IMD), mold damage (MD), purple mottled and stained (PMS), stinkbug damage (SBD), shriveled/wrinkle damage (SWD), and tear damage (TAD). Due to on-field collection a high degree of imbalance was encountered among the damage classes with healthy being the top-class accounting for the 41% of the total dataset while SBD and PMS being the classes with least number of images; accounting for just 5% of total dataset. Secondly, three advanced memory-efficient Deep-CNN models, namely, EfficientNet-B0, ResNet- 50, and VGG- 16, are utilized and fine-tunned to classify damaged soybean kernels. Results from experiments demonstrate that the EfficientNet-B0 model outperforms others in terms of accuracy, average recall, and F1-score and second best in terms of precision. The individual class accuracy achieved is as follows: 77% for HD class, 92% for healthy class, 78% for IMD class, 77% for MD class, 84% for PMS class, 72% for SBD class, 75% for SWD class and 92% for TAD class. In addition, the performance of model in handling of class imbalance among the eight damage classes is also analyzed by comparing the F1-score. Five out of eight classes achieved a F1-score of above 80% including the PMS. The class having the least F1-score was SBD with a score of 68%. The EfficientNet-B0 model attains an overall classification accuracy of 85% with a nominal size of 47 MB. It also has a minimum prediction time of under 9 s while predicting 1480 data points simultaneously. In summary, this study shows that using Deep CNN architectures on a high-magnified and highly unbalanced complex image dataset can accurately classify damaged soybean kernels. The model also performs well in handling data imbalance, making it a useful tool for objective quality assessment of damaged soybean grains in market and trading locations.</p></div>","PeriodicalId":631,"journal":{"name":"Journal of Food Measurement and Characterization","volume":"19 5","pages":"3471 - 3495"},"PeriodicalIF":2.9000,"publicationDate":"2025-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Food Measurement and Characterization","FirstCategoryId":"97","ListUrlMain":"https://link.springer.com/article/10.1007/s11694-025-03195-9","RegionNum":3,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"FOOD SCIENCE & TECHNOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

The assessment of the surface quality of pre-processed soybean kernels is crucial in determining their market acceptance, storage stability, processing quality, and overall consumer approval. Conventional techniques of surface quality evaluation are time-consuming, reliant on personal judgement, and lack consistency. Conversely, the existing techniques are restricted to either selecting healthy soybean kernels from damaged ones without categorizing the damaged ones, or separating different varieties. The lack of a labelled, high-magnification image dataset and the use of advanced CNN models have hindered the exploration of a detailed classification of damage in soybean kernels. These models excel at end-to-end tasks, minimize pre-processing, and eliminate the need for manual feature extraction, enabling quick, accurate, and precise classification. This study demonstrates the use of a machine vision system to create an image dataset consisting of 9866 high-magnification (2.85 µm/pixel) images of soybean kernels with damages. The dataset encompasses eight distinct damage classes: healthy, heat damage (HD), immature damage (IMD), mold damage (MD), purple mottled and stained (PMS), stinkbug damage (SBD), shriveled/wrinkle damage (SWD), and tear damage (TAD). Due to on-field collection a high degree of imbalance was encountered among the damage classes with healthy being the top-class accounting for the 41% of the total dataset while SBD and PMS being the classes with least number of images; accounting for just 5% of total dataset. Secondly, three advanced memory-efficient Deep-CNN models, namely, EfficientNet-B0, ResNet- 50, and VGG- 16, are utilized and fine-tunned to classify damaged soybean kernels. Results from experiments demonstrate that the EfficientNet-B0 model outperforms others in terms of accuracy, average recall, and F1-score and second best in terms of precision. The individual class accuracy achieved is as follows: 77% for HD class, 92% for healthy class, 78% for IMD class, 77% for MD class, 84% for PMS class, 72% for SBD class, 75% for SWD class and 92% for TAD class. In addition, the performance of model in handling of class imbalance among the eight damage classes is also analyzed by comparing the F1-score. Five out of eight classes achieved a F1-score of above 80% including the PMS. The class having the least F1-score was SBD with a score of 68%. The EfficientNet-B0 model attains an overall classification accuracy of 85% with a nominal size of 47 MB. It also has a minimum prediction time of under 9 s while predicting 1480 data points simultaneously. In summary, this study shows that using Deep CNN architectures on a high-magnified and highly unbalanced complex image dataset can accurately classify damaged soybean kernels. The model also performs well in handling data imbalance, making it a useful tool for objective quality assessment of damaged soybean grains in market and trading locations.

基于cnn的大豆籽粒损伤分类研究
预加工大豆仁的表面质量评估对于决定其市场接受度、储存稳定性、加工质量和整体消费者认可至关重要。传统的表面质量评价方法耗时长,依赖于个人判断,缺乏一致性。相反,现有的技术要么局限于从受损的大豆籽粒中选择健康的大豆籽粒,而不对受损的大豆籽粒进行分类,要么局限于区分不同的品种。由于缺乏标记的、高倍率的图像数据集和使用先进的CNN模型,阻碍了对大豆籽粒损伤的详细分类的探索。这些模型擅长端到端任务,最大限度地减少了预处理,消除了手动特征提取的需要,实现了快速、准确和精确的分类。本研究演示了使用机器视觉系统创建一个图像数据集,该数据集由9866张高倍率(2.85µm/pixel)的受损大豆仁图像组成。该数据集包括八种不同的损伤类别:健康、热损伤(HD)、未成熟损伤(IMD)、霉菌损伤(MD)、紫色斑驳和染色(PMS)、臭虫损伤(SBD)、萎缩/皱纹损伤(SWD)和撕裂损伤(TAD)。由于现场采集,各损伤类别之间存在高度的不平衡,其中健康是最高类别,占总数据集的41%,而SBD和PMS是图像数量最少的类别;仅占总数据集的5%。其次,利用三种先进的高效内存深度cnn模型(EfficientNet-B0、ResNet- 50和VGG- 16)对受损大豆籽粒进行分类和微调;实验结果表明,EfficientNet-B0模型在准确率、平均召回率和f1得分方面优于其他模型,在准确率方面排名第二。实现的个体分类准确率如下:HD类77%,健康类92%,IMD类78%,MD类77%,PMS类84%,SBD类72%,SWD类75%,TAD类92%。此外,通过比较f1得分,分析了模型在处理8种伤害类别不平衡方面的表现。包括PMS在内,8个班级中有5个班级的f1分超过80%。f1得分最低的班级是SBD,得分为68%。在47mb的标称大小下,effentnet - b0模型的总体分类准确率达到85%,同时预测1480个数据点的最小预测时间小于9 s。综上所述,本研究表明,在高度放大和高度不平衡的复杂图像数据集上使用深度CNN架构可以准确地对受损大豆仁进行分类。该模型在处理数据不平衡方面也表现良好,为市场和交易场所对受损大豆籽粒进行客观质量评价提供了有效工具。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Food Measurement and Characterization
Journal of Food Measurement and Characterization Agricultural and Biological Sciences-Food Science
CiteScore
6.00
自引率
11.80%
发文量
425
期刊介绍: This interdisciplinary journal publishes new measurement results, characteristic properties, differentiating patterns, measurement methods and procedures for such purposes as food process innovation, product development, quality control, and safety assurance. The journal encompasses all topics related to food property measurement and characterization, including all types of measured properties of food and food materials, features and patterns, measurement principles and techniques, development and evaluation of technologies, novel uses and applications, and industrial implementation of systems and procedures.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信