Minghao Wang, Shaoyi Du, Juan Wang, Hongcheng Han, Huanhuan Huo, Dong Zhang, Shanshan Yu, Jue Jiang
{"title":"基于双向时空背景融合的经食管超声心动图分段任何模型","authors":"Minghao Wang, Shaoyi Du, Juan Wang, Hongcheng Han, Huanhuan Huo, Dong Zhang, Shanshan Yu, Jue Jiang","doi":"10.1016/j.inffus.2025.103771","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate segmentation of the left atrial appendage (LAA) in transesophageal echocardiography is the foundation for clinical evaluation. However, the ambiguous boundaries of the LAA, together with ultrasound noise and complex cardiac motion, make it challenging to obtain temporally consistent and spatially reliable segmentation results. Furthermore, existing works often process spatial and temporal features in isolation, without effectively leveraging spatiotemporal context fusion to enhance segmentation performance. To address these challenges, we propose a Segment Anything Model Based on Bidirectional Spatiotemporal Context Fusion (BiSTC-SAM). First, we design a spatiotemporal context network that encodes effective pixels associated with target changes, thereby mining temporal cues from spatial features. Building on this, we further develop a multi-scale context memory network that performs dynamic feature alignment, thereby integrating temporal representations to refine spatial features. We evaluate the segmentation and generalization performance of our method on a self-constructed transesophageal echocardiography dataset, and further assess its adaptability to different modalities on two publicly available transthoracic echocardiography datasets. Experimental results demonstrate that our method outperforms competing methods in terms of boundary segmentation accuracy and temporal consistency.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103771"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A segment anything model for transesophageal echocardiography based on bidirectional spatiotemporal context fusion\",\"authors\":\"Minghao Wang, Shaoyi Du, Juan Wang, Hongcheng Han, Huanhuan Huo, Dong Zhang, Shanshan Yu, Jue Jiang\",\"doi\":\"10.1016/j.inffus.2025.103771\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Accurate segmentation of the left atrial appendage (LAA) in transesophageal echocardiography is the foundation for clinical evaluation. However, the ambiguous boundaries of the LAA, together with ultrasound noise and complex cardiac motion, make it challenging to obtain temporally consistent and spatially reliable segmentation results. Furthermore, existing works often process spatial and temporal features in isolation, without effectively leveraging spatiotemporal context fusion to enhance segmentation performance. To address these challenges, we propose a Segment Anything Model Based on Bidirectional Spatiotemporal Context Fusion (BiSTC-SAM). First, we design a spatiotemporal context network that encodes effective pixels associated with target changes, thereby mining temporal cues from spatial features. Building on this, we further develop a multi-scale context memory network that performs dynamic feature alignment, thereby integrating temporal representations to refine spatial features. We evaluate the segmentation and generalization performance of our method on a self-constructed transesophageal echocardiography dataset, and further assess its adaptability to different modalities on two publicly available transthoracic echocardiography datasets. Experimental results demonstrate that our method outperforms competing methods in terms of boundary segmentation accuracy and temporal consistency.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103771\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-09-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525008334\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525008334","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
A segment anything model for transesophageal echocardiography based on bidirectional spatiotemporal context fusion
Accurate segmentation of the left atrial appendage (LAA) in transesophageal echocardiography is the foundation for clinical evaluation. However, the ambiguous boundaries of the LAA, together with ultrasound noise and complex cardiac motion, make it challenging to obtain temporally consistent and spatially reliable segmentation results. Furthermore, existing works often process spatial and temporal features in isolation, without effectively leveraging spatiotemporal context fusion to enhance segmentation performance. To address these challenges, we propose a Segment Anything Model Based on Bidirectional Spatiotemporal Context Fusion (BiSTC-SAM). First, we design a spatiotemporal context network that encodes effective pixels associated with target changes, thereby mining temporal cues from spatial features. Building on this, we further develop a multi-scale context memory network that performs dynamic feature alignment, thereby integrating temporal representations to refine spatial features. We evaluate the segmentation and generalization performance of our method on a self-constructed transesophageal echocardiography dataset, and further assess its adaptability to different modalities on two publicly available transthoracic echocardiography datasets. Experimental results demonstrate that our method outperforms competing methods in terms of boundary segmentation accuracy and temporal consistency.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.