Mujtaba Asad , Waqar Azeem , Aftab Ahmad Malik , He Jiang , Ahmad Ali , Jie Yang , Wei Liu
{"title":"3D-MMFN: Multi-level multimodal fusion network for 3D industrial image anomaly detection","authors":"Mujtaba Asad , Waqar Azeem , Aftab Ahmad Malik , He Jiang , Ahmad Ali , Jie Yang , Wei Liu","doi":"10.1016/j.aei.2025.103284","DOIUrl":null,"url":null,"abstract":"<div><div>3D-based image anomaly detection (AD) is a crucial computer vision task in industrial manufacturing. Most existing methods only focus on 2D shape-based detections. However, there is still limited research for detecting anomalies in 3D shapes using multimodal features. Some existing techniques developed for this task are mostly unsuitable for industrial defect detection for several reasons. Firstly, they rely mostly on memory banks, resulting in high storage overheads, making them difficult to deploy on production lines. Secondly, the multimodal features, in the existing 3D industrial AD algorithms, are concatenated directly which cause a significant disruption between the features and degrades the detection efficiency. Thirdly, their inference speed is not fast enough to achieve real-time detection. To address these challenges, we propose a deployment-friendly network named 3D-MMFN. Our model comprises of the following components: (1) The pre-trained feature extractors that generate local features from multi-stream inputs of RGB images, surface normal maps, and point clouds. (2) A novel point-to-pixel based fusion module that efficiently fuses multi-level multimodal features to mitigate disruption during the fusion operation. (3) An anomaly generator module that generates anomalous features from normal multimodal fused features, enabling self-supervised training of 3D-MMFN while eliminating the need for extensive memory banks. Experimental results on the MVTec3D-AD and Eyecandies dataset demonstrate the effectiveness of our proposed model, showcasing significant performance improvements over state-of-the-art methods.</div></div>","PeriodicalId":50941,"journal":{"name":"Advanced Engineering Informatics","volume":"65 ","pages":"Article 103284"},"PeriodicalIF":8.0000,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Engineering Informatics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1474034625001776","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
3D-based image anomaly detection (AD) is a crucial computer vision task in industrial manufacturing. Most existing methods only focus on 2D shape-based detections. However, there is still limited research for detecting anomalies in 3D shapes using multimodal features. Some existing techniques developed for this task are mostly unsuitable for industrial defect detection for several reasons. Firstly, they rely mostly on memory banks, resulting in high storage overheads, making them difficult to deploy on production lines. Secondly, the multimodal features, in the existing 3D industrial AD algorithms, are concatenated directly which cause a significant disruption between the features and degrades the detection efficiency. Thirdly, their inference speed is not fast enough to achieve real-time detection. To address these challenges, we propose a deployment-friendly network named 3D-MMFN. Our model comprises of the following components: (1) The pre-trained feature extractors that generate local features from multi-stream inputs of RGB images, surface normal maps, and point clouds. (2) A novel point-to-pixel based fusion module that efficiently fuses multi-level multimodal features to mitigate disruption during the fusion operation. (3) An anomaly generator module that generates anomalous features from normal multimodal fused features, enabling self-supervised training of 3D-MMFN while eliminating the need for extensive memory banks. Experimental results on the MVTec3D-AD and Eyecandies dataset demonstrate the effectiveness of our proposed model, showcasing significant performance improvements over state-of-the-art methods.
期刊介绍:
Advanced Engineering Informatics is an international Journal that solicits research papers with an emphasis on 'knowledge' and 'engineering applications'. The Journal seeks original papers that report progress in applying methods of engineering informatics. These papers should have engineering relevance and help provide a scientific base for more reliable, spontaneous, and creative engineering decision-making. Additionally, papers should demonstrate the science of supporting knowledge-intensive engineering tasks and validate the generality, power, and scalability of new methods through rigorous evaluation, preferably both qualitatively and quantitatively. Abstracting and indexing for Advanced Engineering Informatics include Science Citation Index Expanded, Scopus and INSPEC.