Jiaxin Wang, Weifang Zhu, Dehui Xiang, Xinjian Chen, Tao Peng, Qing Peng, Meng Wang, Fei Shi
{"title":"不确定性引导下视网膜OCT图像分割的交叉融合网络","authors":"Jiaxin Wang, Weifang Zhu, Dehui Xiang, Xinjian Chen, Tao Peng, Qing Peng, Meng Wang, Fei Shi","doi":"10.1002/mp.18102","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Background</h3>\n \n <p>Deep learning-based segmentation methods for optical coherence tomography (OCT) have demonstrated outstanding performance. However, the stochastic distribution of training data and the inherent limitations of deep neural networks introduce uncertainty into the segmentation process. Accurately estimating this uncertainty is essential for generating reliable confidence assessments and improving model predictions.</p>\n </section>\n \n <section>\n \n <h3> Purpose</h3>\n \n <p>To address these challenges, we propose a novel uncertainty-guided cross-layer fusion network (UGCFNet) for retinal OCT segmentation. UGCFNet integrates uncertainty quantification into the training process of deep neural networks and leverages this uncertainty to enhance segmentation accuracy.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>Our model employs an encoder–decoder architecture that quantitatively assesses uncertainty at multiple stages, directing the network's focus toward regions with higher uncertainty. By facilitating cross-layer feature fusion, UGCFNet enhances the comprehensive understanding of both semantic information and morphological details. Additionally, we incorporate an improved Bayesian neural network loss function alongside an uncertainty-aware loss function, enabling the network to effectively utilize these mechanisms for better uncertainty modeling.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>We conducted extensive experiments on the publicly available AI-Challenger and OIMHS OCT segmentation datasets. The training, validation, and testing sets of the AI-Challenger dataset are comprised of 32, 8, and 43 OCT volumes, yielding a total of 4096, 1024, and 5504 B-scans, respectively. The training, validation, and testing sets of the OIMHS dataset consist of 100, 25, and 25 OCT volumes, resulting in 2,310, 798, and 751 B-scans, respectively. The results demonstrate that UGCFNet achieves state-of-the-art performance, with average Dice similarity coefficients of 79.47% and 93.22% on the respective datasets.</p>\n </section>\n \n <section>\n \n <h3> Conclusion</h3>\n \n <p>Our proposed UGCFNet significantly advances retinal OCT segmentation by integrating uncertainty guidance and cross-level feature fusion, offering more reliable and accurate segmentation outcomes.</p>\n </section>\n </div>","PeriodicalId":18384,"journal":{"name":"Medical physics","volume":"52 9","pages":""},"PeriodicalIF":3.2000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Uncertainty-guided cross-level fusion network for retinal OCT image segmentation\",\"authors\":\"Jiaxin Wang, Weifang Zhu, Dehui Xiang, Xinjian Chen, Tao Peng, Qing Peng, Meng Wang, Fei Shi\",\"doi\":\"10.1002/mp.18102\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n \\n <section>\\n \\n <h3> Background</h3>\\n \\n <p>Deep learning-based segmentation methods for optical coherence tomography (OCT) have demonstrated outstanding performance. However, the stochastic distribution of training data and the inherent limitations of deep neural networks introduce uncertainty into the segmentation process. Accurately estimating this uncertainty is essential for generating reliable confidence assessments and improving model predictions.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Purpose</h3>\\n \\n <p>To address these challenges, we propose a novel uncertainty-guided cross-layer fusion network (UGCFNet) for retinal OCT segmentation. UGCFNet integrates uncertainty quantification into the training process of deep neural networks and leverages this uncertainty to enhance segmentation accuracy.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Methods</h3>\\n \\n <p>Our model employs an encoder–decoder architecture that quantitatively assesses uncertainty at multiple stages, directing the network's focus toward regions with higher uncertainty. By facilitating cross-layer feature fusion, UGCFNet enhances the comprehensive understanding of both semantic information and morphological details. Additionally, we incorporate an improved Bayesian neural network loss function alongside an uncertainty-aware loss function, enabling the network to effectively utilize these mechanisms for better uncertainty modeling.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Results</h3>\\n \\n <p>We conducted extensive experiments on the publicly available AI-Challenger and OIMHS OCT segmentation datasets. The training, validation, and testing sets of the AI-Challenger dataset are comprised of 32, 8, and 43 OCT volumes, yielding a total of 4096, 1024, and 5504 B-scans, respectively. The training, validation, and testing sets of the OIMHS dataset consist of 100, 25, and 25 OCT volumes, resulting in 2,310, 798, and 751 B-scans, respectively. The results demonstrate that UGCFNet achieves state-of-the-art performance, with average Dice similarity coefficients of 79.47% and 93.22% on the respective datasets.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Conclusion</h3>\\n \\n <p>Our proposed UGCFNet significantly advances retinal OCT segmentation by integrating uncertainty guidance and cross-level feature fusion, offering more reliable and accurate segmentation outcomes.</p>\\n </section>\\n </div>\",\"PeriodicalId\":18384,\"journal\":{\"name\":\"Medical physics\",\"volume\":\"52 9\",\"pages\":\"\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical physics\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://aapm.onlinelibrary.wiley.com/doi/10.1002/mp.18102\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"3","ListUrlMain":"https://aapm.onlinelibrary.wiley.com/doi/10.1002/mp.18102","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
Uncertainty-guided cross-level fusion network for retinal OCT image segmentation
Background
Deep learning-based segmentation methods for optical coherence tomography (OCT) have demonstrated outstanding performance. However, the stochastic distribution of training data and the inherent limitations of deep neural networks introduce uncertainty into the segmentation process. Accurately estimating this uncertainty is essential for generating reliable confidence assessments and improving model predictions.
Purpose
To address these challenges, we propose a novel uncertainty-guided cross-layer fusion network (UGCFNet) for retinal OCT segmentation. UGCFNet integrates uncertainty quantification into the training process of deep neural networks and leverages this uncertainty to enhance segmentation accuracy.
Methods
Our model employs an encoder–decoder architecture that quantitatively assesses uncertainty at multiple stages, directing the network's focus toward regions with higher uncertainty. By facilitating cross-layer feature fusion, UGCFNet enhances the comprehensive understanding of both semantic information and morphological details. Additionally, we incorporate an improved Bayesian neural network loss function alongside an uncertainty-aware loss function, enabling the network to effectively utilize these mechanisms for better uncertainty modeling.
Results
We conducted extensive experiments on the publicly available AI-Challenger and OIMHS OCT segmentation datasets. The training, validation, and testing sets of the AI-Challenger dataset are comprised of 32, 8, and 43 OCT volumes, yielding a total of 4096, 1024, and 5504 B-scans, respectively. The training, validation, and testing sets of the OIMHS dataset consist of 100, 25, and 25 OCT volumes, resulting in 2,310, 798, and 751 B-scans, respectively. The results demonstrate that UGCFNet achieves state-of-the-art performance, with average Dice similarity coefficients of 79.47% and 93.22% on the respective datasets.
Conclusion
Our proposed UGCFNet significantly advances retinal OCT segmentation by integrating uncertainty guidance and cross-level feature fusion, offering more reliable and accurate segmentation outcomes.
期刊介绍:
Medical Physics publishes original, high impact physics, imaging science, and engineering research that advances patient diagnosis and therapy through contributions in 1) Basic science developments with high potential for clinical translation 2) Clinical applications of cutting edge engineering and physics innovations 3) Broadly applicable and innovative clinical physics developments
Medical Physics is a journal of global scope and reach. By publishing in Medical Physics your research will reach an international, multidisciplinary audience including practicing medical physicists as well as physics- and engineering based translational scientists. We work closely with authors of promising articles to improve their quality.