Trimming the Risk: Towards Reliable Continuous Training for Deep Learning Inspection Systems

Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid
{"title":"Trimming the Risk: Towards Reliable Continuous Training for Deep Learning Inspection Systems","authors":"Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid","doi":"arxiv-2409.09108","DOIUrl":null,"url":null,"abstract":"The industry increasingly relies on deep learning (DL) technology for\nmanufacturing inspections, which are challenging to automate with rule-based\nmachine vision algorithms. DL-powered inspection systems derive defect patterns\nfrom labeled images, combining human-like agility with the consistency of a\ncomputerized system. However, finite labeled datasets often fail to encompass\nall natural variations necessitating Continuous Training (CT) to regularly\nadjust their models with recent data. Effective CT requires fresh labeled\nsamples from the original distribution; otherwise, selfgenerated labels can\nlead to silent performance degradation. To mitigate this risk, we develop a\nrobust CT-based maintenance approach that updates DL models using reliable data\nselections through a two-stage filtering process. The initial stage filters out\nlow-confidence predictions, as the model inherently discredits them. The second\nstage uses variational auto-encoders and histograms to generate image\nembeddings that capture latent and pixel characteristics, then rejects the\ninputs of substantially shifted embeddings as drifted data with erroneous\noverconfidence. Then, a fine-tuning of the original DL model is executed on the\nfiltered inputs while validating on a mixture of recent production and original\ndatasets. This strategy mitigates catastrophic forgetting and ensures the model\nadapts effectively to new operational conditions. Evaluations on industrial\ninspection systems for popsicle stick prints and glass bottles using critical\nreal-world datasets showed less than 9% of erroneous self-labeled data are\nretained after filtering and used for fine-tuning, improving model performance\non production data by up to 14% without compromising its results on original\nvalidation data.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09108","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The industry increasingly relies on deep learning (DL) technology for manufacturing inspections, which are challenging to automate with rule-based machine vision algorithms. DL-powered inspection systems derive defect patterns from labeled images, combining human-like agility with the consistency of a computerized system. However, finite labeled datasets often fail to encompass all natural variations necessitating Continuous Training (CT) to regularly adjust their models with recent data. Effective CT requires fresh labeled samples from the original distribution; otherwise, selfgenerated labels can lead to silent performance degradation. To mitigate this risk, we develop a robust CT-based maintenance approach that updates DL models using reliable data selections through a two-stage filtering process. The initial stage filters out low-confidence predictions, as the model inherently discredits them. The second stage uses variational auto-encoders and histograms to generate image embeddings that capture latent and pixel characteristics, then rejects the inputs of substantially shifted embeddings as drifted data with erroneous overconfidence. Then, a fine-tuning of the original DL model is executed on the filtered inputs while validating on a mixture of recent production and original datasets. This strategy mitigates catastrophic forgetting and ensures the model adapts effectively to new operational conditions. Evaluations on industrial inspection systems for popsicle stick prints and glass bottles using critical real-world datasets showed less than 9% of erroneous self-labeled data are retained after filtering and used for fine-tuning, improving model performance on production data by up to 14% without compromising its results on original validation data.
降低风险:为深度学习检测系统提供可靠的持续训练
制造业越来越依赖于深度学习(DL)技术来进行制造检测,而使用基于规则的机器视觉算法来实现自动化具有挑战性。由深度学习驱动的检测系统从标记图像中提取缺陷模式,将人类的敏捷性与计算机化系统的一致性结合起来。然而,有限的标注数据集往往无法涵盖所有自然变化,因此需要进行持续训练(CT),利用最新数据定期调整模型。有效的持续训练需要来自原始分布的新鲜标签样本;否则,自生成的标签会导致无声的性能下降。为了降低这种风险,我们开发了一种基于 CT 的稳健维护方法,通过两阶段过滤过程,使用可靠的数据选择更新 DL 模型。第一阶段过滤掉置信度低的预测,因为模型本身就会否定这些预测。第二阶段使用变异自动编码器和直方图来生成图像嵌入,以捕捉潜在特征和像素特征,然后将大幅偏移的嵌入作为具有错误过高置信度的漂移数据剔除。然后,对过滤后的输入执行原始 DL 模型的微调,同时在最新生产数据集和原始数据集的混合数据上进行验证。这种策略可以减少灾难性遗忘,确保模型有效适应新的运行条件。利用重要的真实数据集对冰棒棍印花和玻璃瓶的工业检测系统进行的评估表明,经过过滤并用于微调后,错误的自标注数据只保留了不到 9%,从而将模型在生产数据上的性能提高了 14%,而不会影响其在原始验证数据上的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信