Visual coating inspection framework via self-labeling and multi-stage deep learning strategies

IF 5.9 2区 工程技术 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Changheon Han, Jiho Lee, Martin B. G. Jun, Sang Won Lee, Huitaek Yun
{"title":"Visual coating inspection framework via self-labeling and multi-stage deep learning strategies","authors":"Changheon Han, Jiho Lee, Martin B. G. Jun, Sang Won Lee, Huitaek Yun","doi":"10.1007/s10845-024-02372-9","DOIUrl":null,"url":null,"abstract":"<p>An instantaneous and precise coating inspection method is imperative to mitigate the risk of flaws, defects, and discrepancies on coated surfaces. While many studies have demonstrated the effectiveness of automated visual inspection (AVI) approaches enhanced by computer vision and deep learning, critical challenges exist for practical applications in the manufacturing domain. Computer vision has proven to be inflexible, demanding sophisticated algorithms for diverse feature extraction. In deep learning, supervised approaches are constrained by the need for annotated datasets, whereas unsupervised methods often result in lower performance. Addressing these challenges, this paper proposes a novel deep learning-based automated visual inspection (AVI) framework designed to minimize the necessity for extensive feature engineering, programming, and manual data annotation in classifying fuel injection nozzles and discerning their coating interfaces from scratch. This proposed framework comprises six integral components: It begins by distinguishing between coated and uncoated nozzles through gray level co-occurrence matrix (GLCM)-based texture analysis and autoencoder (AE)-based classification. This is followed by cropping surface images from uncoated nozzles, and then building an AE model to estimate the coating interface locations on coated nozzles. The next step involves generating autonomously annotated datasets derived from these estimated coating interface locations. Subsequently, a convolutional neural network (CNN)-based detection model is trained to accurately localize the coating interface locations. The final component focuses on enhancing model performance and trustworthiness. This framework demonstrated over 95% accuracy in pinpointing the coating interfaces within the error range of ± 6 pixels and processed at a rate of 7.18 images per second. Additionally, explainable artificial intelligence (XAI) techniques such as t-distributed stochastic neighbor embedding (t-SNE) and the integrated gradient substantiated the reliability of the models.</p>","PeriodicalId":16193,"journal":{"name":"Journal of Intelligent Manufacturing","volume":"44 1","pages":""},"PeriodicalIF":5.9000,"publicationDate":"2024-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Intelligent Manufacturing","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s10845-024-02372-9","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

An instantaneous and precise coating inspection method is imperative to mitigate the risk of flaws, defects, and discrepancies on coated surfaces. While many studies have demonstrated the effectiveness of automated visual inspection (AVI) approaches enhanced by computer vision and deep learning, critical challenges exist for practical applications in the manufacturing domain. Computer vision has proven to be inflexible, demanding sophisticated algorithms for diverse feature extraction. In deep learning, supervised approaches are constrained by the need for annotated datasets, whereas unsupervised methods often result in lower performance. Addressing these challenges, this paper proposes a novel deep learning-based automated visual inspection (AVI) framework designed to minimize the necessity for extensive feature engineering, programming, and manual data annotation in classifying fuel injection nozzles and discerning their coating interfaces from scratch. This proposed framework comprises six integral components: It begins by distinguishing between coated and uncoated nozzles through gray level co-occurrence matrix (GLCM)-based texture analysis and autoencoder (AE)-based classification. This is followed by cropping surface images from uncoated nozzles, and then building an AE model to estimate the coating interface locations on coated nozzles. The next step involves generating autonomously annotated datasets derived from these estimated coating interface locations. Subsequently, a convolutional neural network (CNN)-based detection model is trained to accurately localize the coating interface locations. The final component focuses on enhancing model performance and trustworthiness. This framework demonstrated over 95% accuracy in pinpointing the coating interfaces within the error range of ± 6 pixels and processed at a rate of 7.18 images per second. Additionally, explainable artificial intelligence (XAI) techniques such as t-distributed stochastic neighbor embedding (t-SNE) and the integrated gradient substantiated the reliability of the models.

Abstract Image

通过自标记和多级深度学习策略实现可视化涂层检测框架
要降低涂层表面出现瑕疵、缺陷和差异的风险,就必须采用即时、精确的涂层检测方法。虽然许多研究已经证明了通过计算机视觉和深度学习增强的自动视觉检测(AVI)方法的有效性,但在制造领域的实际应用中仍存在严峻挑战。事实证明,计算机视觉并不灵活,需要复杂的算法来提取各种特征。在深度学习中,有监督的方法受到需要注释数据集的限制,而无监督的方法往往会导致性能降低。为了应对这些挑战,本文提出了一种新颖的基于深度学习的自动视觉检测(AVI)框架,旨在最大限度地减少在对喷油嘴进行分类并从头开始辨别其涂层界面时进行大量特征工程、编程和手动数据注释的必要性。该拟议框架由六个组成部分组成:首先,它通过基于灰度共现矩阵 (GLCM) 的纹理分析和基于自动编码器 (AE) 的分类来区分有涂层和无涂层的喷嘴。随后,对未涂层喷嘴的表面图像进行裁剪,然后建立 AE 模型来估计涂层喷嘴的涂层界面位置。下一步是根据这些估计的涂层界面位置生成自主注释数据集。随后,对基于卷积神经网络 (CNN) 的检测模型进行训练,以准确定位涂层界面位置。最后一个部分的重点是提高模型的性能和可信度。该框架在误差± 6 像素范围内精确定位涂层界面的准确率超过 95%,处理速度为每秒 7.18 幅图像。此外,可解释人工智能(XAI)技术,如 t 分布随机邻域嵌入(t-SNE)和综合梯度,也证实了模型的可靠性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Intelligent Manufacturing
Journal of Intelligent Manufacturing 工程技术-工程:制造
CiteScore
19.30
自引率
9.60%
发文量
171
审稿时长
5.2 months
期刊介绍: The Journal of Nonlinear Engineering aims to be a platform for sharing original research results in theoretical, experimental, practical, and applied nonlinear phenomena within engineering. It serves as a forum to exchange ideas and applications of nonlinear problems across various engineering disciplines. Articles are considered for publication if they explore nonlinearities in engineering systems, offering realistic mathematical modeling, utilizing nonlinearity for new designs, stabilizing systems, understanding system behavior through nonlinearity, optimizing systems based on nonlinear interactions, and developing algorithms to harness and leverage nonlinear elements.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信