Yu Yan , Lei Zhang , Jiayi Li , Leyi Zhang , Zhang Yi
{"title":"FUNet:用于脑肿瘤分割的信道多模态融合及不确定区域调整网络","authors":"Yu Yan , Lei Zhang , Jiayi Li , Leyi Zhang , Zhang Yi","doi":"10.1016/j.inffus.2025.103474","DOIUrl":null,"url":null,"abstract":"<div><div>Multi-modal images are crucial for enhancing the performance of brain tumor segmentation. Existing multi-modal brain tumor segmentation methods have the following three main shortcomings: To begin with, framework design remains underexplored in current research. Secondly, effectively fusing multi-modal data, which characterize brain tumors differently, poses a significant challenge. Finally, uncertain and error-prone regions may exist within the fused features, complicating subsequent analysis. To address these issues, we propose Frequency Channel Multi-Modal Fusion and Uncertain Region Adjustment Network (FUNet). FUNet employs a triple-parallel-stream framework to integrate multi-modal information. In the encoder of the multi-modal information learning stream, we design a frequency channel multi-modal fusion module (FCMM), which distinguishes between the complementarity and redundancy of the modal information and mines the intrinsic connection. Additionally, in the decoder, we design an uncertain region adjustment module (URAM), which generates an adjustment factor to enable pixel-wise adjust uncertain error-prone regions existing in the fused features. Experiments on BrsTS 2018 and BraTS-PED 2023 demonstrate that our method achieves better results than other state-of-the-art methods.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"125 ","pages":"Article 103474"},"PeriodicalIF":15.5000,"publicationDate":"2025-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FUNet: Frequency Channel Multi-Modal Fusion and Uncertain Region Adjustment Network for brain tumor segmentation\",\"authors\":\"Yu Yan , Lei Zhang , Jiayi Li , Leyi Zhang , Zhang Yi\",\"doi\":\"10.1016/j.inffus.2025.103474\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Multi-modal images are crucial for enhancing the performance of brain tumor segmentation. Existing multi-modal brain tumor segmentation methods have the following three main shortcomings: To begin with, framework design remains underexplored in current research. Secondly, effectively fusing multi-modal data, which characterize brain tumors differently, poses a significant challenge. Finally, uncertain and error-prone regions may exist within the fused features, complicating subsequent analysis. To address these issues, we propose Frequency Channel Multi-Modal Fusion and Uncertain Region Adjustment Network (FUNet). FUNet employs a triple-parallel-stream framework to integrate multi-modal information. In the encoder of the multi-modal information learning stream, we design a frequency channel multi-modal fusion module (FCMM), which distinguishes between the complementarity and redundancy of the modal information and mines the intrinsic connection. Additionally, in the decoder, we design an uncertain region adjustment module (URAM), which generates an adjustment factor to enable pixel-wise adjust uncertain error-prone regions existing in the fused features. Experiments on BrsTS 2018 and BraTS-PED 2023 demonstrate that our method achieves better results than other state-of-the-art methods.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"125 \",\"pages\":\"Article 103474\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-07-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525005470\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525005470","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
FUNet: Frequency Channel Multi-Modal Fusion and Uncertain Region Adjustment Network for brain tumor segmentation
Multi-modal images are crucial for enhancing the performance of brain tumor segmentation. Existing multi-modal brain tumor segmentation methods have the following three main shortcomings: To begin with, framework design remains underexplored in current research. Secondly, effectively fusing multi-modal data, which characterize brain tumors differently, poses a significant challenge. Finally, uncertain and error-prone regions may exist within the fused features, complicating subsequent analysis. To address these issues, we propose Frequency Channel Multi-Modal Fusion and Uncertain Region Adjustment Network (FUNet). FUNet employs a triple-parallel-stream framework to integrate multi-modal information. In the encoder of the multi-modal information learning stream, we design a frequency channel multi-modal fusion module (FCMM), which distinguishes between the complementarity and redundancy of the modal information and mines the intrinsic connection. Additionally, in the decoder, we design an uncertain region adjustment module (URAM), which generates an adjustment factor to enable pixel-wise adjust uncertain error-prone regions existing in the fused features. Experiments on BrsTS 2018 and BraTS-PED 2023 demonstrate that our method achieves better results than other state-of-the-art methods.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.