{"title":"When explainable artificial intelligence meets data governance: Enhancing trustworthiness in multimodal gas classification","authors":"Sapdo Utomo , Ayush Pratap , Periyasami Karthikeyan , John Ayeelyan , Hsiu-Chun Hsu , Pao-Ann Hsiung","doi":"10.1016/j.inffus.2025.103440","DOIUrl":null,"url":null,"abstract":"<div><div>In the domain of artificial intelligence, the incorporation of multimodal data has become increasingly popular as a method to improve the performance of models by offering a more comprehensive range of information for learning. The gas classification, which is of utmost importance in multiple fields like industry, security, and healthcare, has received considerable interest. Nevertheless, a review of current research indicates that numerous studies suffer from inadequate data governance, resulting in subpar performance despite how complicated their proposed methodologies are. Although explainable artificial intelligence (XAI) techniques are gaining recognition for their ability to assist researchers in analyzing and enhancing model performance, their use in multimodal gas classification is still limited. This research presents a method that integrates strong data governance practices with XAI to improve the accuracy of models in classifying different types of gases using multimodal input. Our approach enhances data quality and offers a very efficient model architecture with a minimal parameter number of 0.8 million. The proposed model attains testing accuracies of 98.49% for the sensor modality, 96.48% for the image modality, and 99.4% for the fusion modality, surpassing the maximum accuracy achieved by existing state-of-the-art models, which stands at 99.2% with 106 million parameters. Notably, our model is 132.5 times smaller than the most accurate model currently used in multimodal gas classification studies. The proposed model’s robustness and trustworthiness are confirmed by extensive testing. The results demonstrate that our approach makes a significant contribution to the field of multimodal classification.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"125 ","pages":"Article 103440"},"PeriodicalIF":14.7000,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525005135","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
In the domain of artificial intelligence, the incorporation of multimodal data has become increasingly popular as a method to improve the performance of models by offering a more comprehensive range of information for learning. The gas classification, which is of utmost importance in multiple fields like industry, security, and healthcare, has received considerable interest. Nevertheless, a review of current research indicates that numerous studies suffer from inadequate data governance, resulting in subpar performance despite how complicated their proposed methodologies are. Although explainable artificial intelligence (XAI) techniques are gaining recognition for their ability to assist researchers in analyzing and enhancing model performance, their use in multimodal gas classification is still limited. This research presents a method that integrates strong data governance practices with XAI to improve the accuracy of models in classifying different types of gases using multimodal input. Our approach enhances data quality and offers a very efficient model architecture with a minimal parameter number of 0.8 million. The proposed model attains testing accuracies of 98.49% for the sensor modality, 96.48% for the image modality, and 99.4% for the fusion modality, surpassing the maximum accuracy achieved by existing state-of-the-art models, which stands at 99.2% with 106 million parameters. Notably, our model is 132.5 times smaller than the most accurate model currently used in multimodal gas classification studies. The proposed model’s robustness and trustworthiness are confirmed by extensive testing. The results demonstrate that our approach makes a significant contribution to the field of multimodal classification.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.