Jianhong Gan, Runqing Kang, Xun Deng, Tongli He, Nie Yu, Yuling Gan, Peiyang Wei, Xiangyi Chen, Xiaoli Peng, Zhibin Li
{"title":"基于分段任意模型和半监督师生模型的口腔CBCT图像分割方法。","authors":"Jianhong Gan, Runqing Kang, Xun Deng, Tongli He, Nie Yu, Yuling Gan, Peiyang Wei, Xiangyi Chen, Xiaoli Peng, Zhibin Li","doi":"10.1002/mp.17854","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Accurate segmentation of oral cone beam computed tomography (CBCT) images is essential for research and clinical diagnosis. However, irregular and blurred tooth boundaries in CBCT images complicate the labeling of oral tissues, and insufficient labeled samples further limit the generalization ability of segmentation models. The Segment Anything Model (SAM) demonstrates strong generalization and segmentation accuracy across diverse tasks as a vision foundation model. The Teacher-Student (TS) model has proven effective in semi-supervised learning approaches.</p><p><strong>Purpose: </strong>To accurately segment various parts of oral CBCT, such as enamel, pulp, bone, blood vessels, air, etc., an improved segmentation method named SAM-TS is proposed, which combines SAM with the TS model. SAM-TS leverages Low-Rank Adaptation (LoRA) to fine-tune the SAM model on oral CBCT images with fewer parameters.</p><p><strong>Methods: </strong>To efficiently utilize numerous unlabeled images for training models, the LoRA strategy is improved to fine-tune the SAM. The fine-tuned SAM and teacher models collaboratively generate pseudo-labels on unlabeled images, which are filtered and utilized to train the student model. Then, a data augmentation-based Mean Intersection over Union (MIoU) method is proposed to filter out unreliable or spurious pseudo-labels. Finally, the Exponential Moving Average (EMA) method is used to transfer the student model's parameters to the teacher model. After repeating this process, the final optimized student model for segmentation is obtained. The experimental results demonstrate that incorporating unlabeled data into model training through SAM-TS significantly enhances the model's generalization ability and segmentation accuracy.</p><p><strong>Results: </strong>Compared to the baseline algorithm, the proposed method achieves an overall improvement of over 6.48% in MIoU. In the tooth segmentation task, the minimum MIoU and maximum MIoU increased by at least 10% and 27.32%, respectively. In the bone segmentation task, the minimum MIoU and maximum MIoU increased by 7.9% and 32.44%, respectively. Additionally, for overall segmentation, the Hausdorff distance (HD) decreased by 5.1 mm, and the Dice coefficient increased by 2.87%.</p><p><strong>Conclusion: </strong>SAM-TS outperforms existing semi-supervised methods, offering a more competitive and efficient approach to CBCT image segmentation. This method addresses the data annotation bottleneck and opens new avenues for semi-supervised learning applications in medical imaging.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A segmentation method for oral CBCT image based on Segment Anything Model and semi-supervised teacher-student model.\",\"authors\":\"Jianhong Gan, Runqing Kang, Xun Deng, Tongli He, Nie Yu, Yuling Gan, Peiyang Wei, Xiangyi Chen, Xiaoli Peng, Zhibin Li\",\"doi\":\"10.1002/mp.17854\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Accurate segmentation of oral cone beam computed tomography (CBCT) images is essential for research and clinical diagnosis. However, irregular and blurred tooth boundaries in CBCT images complicate the labeling of oral tissues, and insufficient labeled samples further limit the generalization ability of segmentation models. The Segment Anything Model (SAM) demonstrates strong generalization and segmentation accuracy across diverse tasks as a vision foundation model. The Teacher-Student (TS) model has proven effective in semi-supervised learning approaches.</p><p><strong>Purpose: </strong>To accurately segment various parts of oral CBCT, such as enamel, pulp, bone, blood vessels, air, etc., an improved segmentation method named SAM-TS is proposed, which combines SAM with the TS model. SAM-TS leverages Low-Rank Adaptation (LoRA) to fine-tune the SAM model on oral CBCT images with fewer parameters.</p><p><strong>Methods: </strong>To efficiently utilize numerous unlabeled images for training models, the LoRA strategy is improved to fine-tune the SAM. The fine-tuned SAM and teacher models collaboratively generate pseudo-labels on unlabeled images, which are filtered and utilized to train the student model. Then, a data augmentation-based Mean Intersection over Union (MIoU) method is proposed to filter out unreliable or spurious pseudo-labels. Finally, the Exponential Moving Average (EMA) method is used to transfer the student model's parameters to the teacher model. After repeating this process, the final optimized student model for segmentation is obtained. The experimental results demonstrate that incorporating unlabeled data into model training through SAM-TS significantly enhances the model's generalization ability and segmentation accuracy.</p><p><strong>Results: </strong>Compared to the baseline algorithm, the proposed method achieves an overall improvement of over 6.48% in MIoU. In the tooth segmentation task, the minimum MIoU and maximum MIoU increased by at least 10% and 27.32%, respectively. In the bone segmentation task, the minimum MIoU and maximum MIoU increased by 7.9% and 32.44%, respectively. Additionally, for overall segmentation, the Hausdorff distance (HD) decreased by 5.1 mm, and the Dice coefficient increased by 2.87%.</p><p><strong>Conclusion: </strong>SAM-TS outperforms existing semi-supervised methods, offering a more competitive and efficient approach to CBCT image segmentation. This method addresses the data annotation bottleneck and opens new avenues for semi-supervised learning applications in medical imaging.</p>\",\"PeriodicalId\":94136,\"journal\":{\"name\":\"Medical physics\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-05-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical physics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1002/mp.17854\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/mp.17854","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A segmentation method for oral CBCT image based on Segment Anything Model and semi-supervised teacher-student model.
Background: Accurate segmentation of oral cone beam computed tomography (CBCT) images is essential for research and clinical diagnosis. However, irregular and blurred tooth boundaries in CBCT images complicate the labeling of oral tissues, and insufficient labeled samples further limit the generalization ability of segmentation models. The Segment Anything Model (SAM) demonstrates strong generalization and segmentation accuracy across diverse tasks as a vision foundation model. The Teacher-Student (TS) model has proven effective in semi-supervised learning approaches.
Purpose: To accurately segment various parts of oral CBCT, such as enamel, pulp, bone, blood vessels, air, etc., an improved segmentation method named SAM-TS is proposed, which combines SAM with the TS model. SAM-TS leverages Low-Rank Adaptation (LoRA) to fine-tune the SAM model on oral CBCT images with fewer parameters.
Methods: To efficiently utilize numerous unlabeled images for training models, the LoRA strategy is improved to fine-tune the SAM. The fine-tuned SAM and teacher models collaboratively generate pseudo-labels on unlabeled images, which are filtered and utilized to train the student model. Then, a data augmentation-based Mean Intersection over Union (MIoU) method is proposed to filter out unreliable or spurious pseudo-labels. Finally, the Exponential Moving Average (EMA) method is used to transfer the student model's parameters to the teacher model. After repeating this process, the final optimized student model for segmentation is obtained. The experimental results demonstrate that incorporating unlabeled data into model training through SAM-TS significantly enhances the model's generalization ability and segmentation accuracy.
Results: Compared to the baseline algorithm, the proposed method achieves an overall improvement of over 6.48% in MIoU. In the tooth segmentation task, the minimum MIoU and maximum MIoU increased by at least 10% and 27.32%, respectively. In the bone segmentation task, the minimum MIoU and maximum MIoU increased by 7.9% and 32.44%, respectively. Additionally, for overall segmentation, the Hausdorff distance (HD) decreased by 5.1 mm, and the Dice coefficient increased by 2.87%.
Conclusion: SAM-TS outperforms existing semi-supervised methods, offering a more competitive and efficient approach to CBCT image segmentation. This method addresses the data annotation bottleneck and opens new avenues for semi-supervised learning applications in medical imaging.