Shijie Zhang , Shaozheng He , Jingjing Wu , Dandan Wang , Pan Zeng , Guorong Lyu
{"title":"一个多特征融合模型测量胎儿在分娩过程中的头部屈曲变形和变形多头自注意","authors":"Shijie Zhang , Shaozheng He , Jingjing Wu , Dandan Wang , Pan Zeng , Guorong Lyu","doi":"10.1016/j.compmedimag.2025.102582","DOIUrl":null,"url":null,"abstract":"<div><div>Fetal head flexion is essential during labor. The current assessment presents technical challenges for unskilled ultrasound operators. Therefore, the study aimed to propose an occiput-spine angle measurement network (OSAM-NET) to improve the accuracy and intrapartum applicability of fetal head flexion assessment. We used YOLOv8 to extract key anatomical structures (occiput and spine) and Vision Transformer to extract global features on ultrasound images of the posterior mid-sagittal section of the fetus (PMSF). Then, by fusing the geometric location information and global features through the multi-head self-attention mechanism, we constructed the multi-feature fusion model OSAM-NET. The model was able to extract intricate features from multi-dimensional information, effectively boosting the accuracy of occiput-spine angle (OSA) prediction. We curated the first OSA dataset comprising 1688 high-quality, clear PMSF ultrasound images and annotations to train and test our model. We evaluated the performance of OSAM-NET and compared it with other models. The results showed that OSAM-NET outperformed the comparison models on all evaluation metrics, with R² increasing by nearly 13 %, and root mean square error (RMSE) and mean absolute error (MAE) decreasing by approximately 15 % and 20 %, respectively. The intraclass correlation coefficient (ICC) improved by about 8 %, with the average value reaching 0.89, indicating good agreement with the measurements of ultrasound experts. The multi-feature fusion model OSAM-NET demonstrates strong applicability and predictive accuracy for complex ultrasound images. This study provides a reliable and efficient tool for automatically evaluating fetal head flexion. The real-world application potential has been validated on prospective test dataset.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"124 ","pages":"Article 102582"},"PeriodicalIF":5.4000,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"OSAM-NET: A multi-feature fusion model for measuring fetal head flexion during labor with transformer multi-head self-attention\",\"authors\":\"Shijie Zhang , Shaozheng He , Jingjing Wu , Dandan Wang , Pan Zeng , Guorong Lyu\",\"doi\":\"10.1016/j.compmedimag.2025.102582\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Fetal head flexion is essential during labor. The current assessment presents technical challenges for unskilled ultrasound operators. Therefore, the study aimed to propose an occiput-spine angle measurement network (OSAM-NET) to improve the accuracy and intrapartum applicability of fetal head flexion assessment. We used YOLOv8 to extract key anatomical structures (occiput and spine) and Vision Transformer to extract global features on ultrasound images of the posterior mid-sagittal section of the fetus (PMSF). Then, by fusing the geometric location information and global features through the multi-head self-attention mechanism, we constructed the multi-feature fusion model OSAM-NET. The model was able to extract intricate features from multi-dimensional information, effectively boosting the accuracy of occiput-spine angle (OSA) prediction. We curated the first OSA dataset comprising 1688 high-quality, clear PMSF ultrasound images and annotations to train and test our model. We evaluated the performance of OSAM-NET and compared it with other models. The results showed that OSAM-NET outperformed the comparison models on all evaluation metrics, with R² increasing by nearly 13 %, and root mean square error (RMSE) and mean absolute error (MAE) decreasing by approximately 15 % and 20 %, respectively. The intraclass correlation coefficient (ICC) improved by about 8 %, with the average value reaching 0.89, indicating good agreement with the measurements of ultrasound experts. The multi-feature fusion model OSAM-NET demonstrates strong applicability and predictive accuracy for complex ultrasound images. This study provides a reliable and efficient tool for automatically evaluating fetal head flexion. The real-world application potential has been validated on prospective test dataset.</div></div>\",\"PeriodicalId\":50631,\"journal\":{\"name\":\"Computerized Medical Imaging and Graphics\",\"volume\":\"124 \",\"pages\":\"Article 102582\"},\"PeriodicalIF\":5.4000,\"publicationDate\":\"2025-05-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computerized Medical Imaging and Graphics\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0895611125000916\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computerized Medical Imaging and Graphics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0895611125000916","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
OSAM-NET: A multi-feature fusion model for measuring fetal head flexion during labor with transformer multi-head self-attention
Fetal head flexion is essential during labor. The current assessment presents technical challenges for unskilled ultrasound operators. Therefore, the study aimed to propose an occiput-spine angle measurement network (OSAM-NET) to improve the accuracy and intrapartum applicability of fetal head flexion assessment. We used YOLOv8 to extract key anatomical structures (occiput and spine) and Vision Transformer to extract global features on ultrasound images of the posterior mid-sagittal section of the fetus (PMSF). Then, by fusing the geometric location information and global features through the multi-head self-attention mechanism, we constructed the multi-feature fusion model OSAM-NET. The model was able to extract intricate features from multi-dimensional information, effectively boosting the accuracy of occiput-spine angle (OSA) prediction. We curated the first OSA dataset comprising 1688 high-quality, clear PMSF ultrasound images and annotations to train and test our model. We evaluated the performance of OSAM-NET and compared it with other models. The results showed that OSAM-NET outperformed the comparison models on all evaluation metrics, with R² increasing by nearly 13 %, and root mean square error (RMSE) and mean absolute error (MAE) decreasing by approximately 15 % and 20 %, respectively. The intraclass correlation coefficient (ICC) improved by about 8 %, with the average value reaching 0.89, indicating good agreement with the measurements of ultrasound experts. The multi-feature fusion model OSAM-NET demonstrates strong applicability and predictive accuracy for complex ultrasound images. This study provides a reliable and efficient tool for automatically evaluating fetal head flexion. The real-world application potential has been validated on prospective test dataset.
期刊介绍:
The purpose of the journal Computerized Medical Imaging and Graphics is to act as a source for the exchange of research results concerning algorithmic advances, development, and application of digital imaging in disease detection, diagnosis, intervention, prevention, precision medicine, and population health. Included in the journal will be articles on novel computerized imaging or visualization techniques, including artificial intelligence and machine learning, augmented reality for surgical planning and guidance, big biomedical data visualization, computer-aided diagnosis, computerized-robotic surgery, image-guided therapy, imaging scanning and reconstruction, mobile and tele-imaging, radiomics, and imaging integration and modeling with other information relevant to digital health. The types of biomedical imaging include: magnetic resonance, computed tomography, ultrasound, nuclear medicine, X-ray, microwave, optical and multi-photon microscopy, video and sensory imaging, and the convergence of biomedical images with other non-imaging datasets.