Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid
{"title":"Trimming the Risk: Towards Reliable Continuous Training for Deep Learning Inspection Systems","authors":"Altaf Allah Abbassi, Houssem Ben Braiek, Foutse Khomh, Thomas Reid","doi":"arxiv-2409.09108","DOIUrl":null,"url":null,"abstract":"The industry increasingly relies on deep learning (DL) technology for\nmanufacturing inspections, which are challenging to automate with rule-based\nmachine vision algorithms. DL-powered inspection systems derive defect patterns\nfrom labeled images, combining human-like agility with the consistency of a\ncomputerized system. However, finite labeled datasets often fail to encompass\nall natural variations necessitating Continuous Training (CT) to regularly\nadjust their models with recent data. Effective CT requires fresh labeled\nsamples from the original distribution; otherwise, selfgenerated labels can\nlead to silent performance degradation. To mitigate this risk, we develop a\nrobust CT-based maintenance approach that updates DL models using reliable data\nselections through a two-stage filtering process. The initial stage filters out\nlow-confidence predictions, as the model inherently discredits them. The second\nstage uses variational auto-encoders and histograms to generate image\nembeddings that capture latent and pixel characteristics, then rejects the\ninputs of substantially shifted embeddings as drifted data with erroneous\noverconfidence. Then, a fine-tuning of the original DL model is executed on the\nfiltered inputs while validating on a mixture of recent production and original\ndatasets. This strategy mitigates catastrophic forgetting and ensures the model\nadapts effectively to new operational conditions. Evaluations on industrial\ninspection systems for popsicle stick prints and glass bottles using critical\nreal-world datasets showed less than 9% of erroneous self-labeled data are\nretained after filtering and used for fine-tuning, improving model performance\non production data by up to 14% without compromising its results on original\nvalidation data.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09108","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The industry increasingly relies on deep learning (DL) technology for
manufacturing inspections, which are challenging to automate with rule-based
machine vision algorithms. DL-powered inspection systems derive defect patterns
from labeled images, combining human-like agility with the consistency of a
computerized system. However, finite labeled datasets often fail to encompass
all natural variations necessitating Continuous Training (CT) to regularly
adjust their models with recent data. Effective CT requires fresh labeled
samples from the original distribution; otherwise, selfgenerated labels can
lead to silent performance degradation. To mitigate this risk, we develop a
robust CT-based maintenance approach that updates DL models using reliable data
selections through a two-stage filtering process. The initial stage filters out
low-confidence predictions, as the model inherently discredits them. The second
stage uses variational auto-encoders and histograms to generate image
embeddings that capture latent and pixel characteristics, then rejects the
inputs of substantially shifted embeddings as drifted data with erroneous
overconfidence. Then, a fine-tuning of the original DL model is executed on the
filtered inputs while validating on a mixture of recent production and original
datasets. This strategy mitigates catastrophic forgetting and ensures the model
adapts effectively to new operational conditions. Evaluations on industrial
inspection systems for popsicle stick prints and glass bottles using critical
real-world datasets showed less than 9% of erroneous self-labeled data are
retained after filtering and used for fine-tuning, improving model performance
on production data by up to 14% without compromising its results on original
validation data.