A segmentation method for oral CBCT image based on Segment Anything Model and semi-supervised teacher-student model.

Medical physics Pub Date : 2025-05-07 DOI:10.1002/mp.17854
Jianhong Gan, Runqing Kang, Xun Deng, Tongli He, Nie Yu, Yuling Gan, Peiyang Wei, Xiangyi Chen, Xiaoli Peng, Zhibin Li
{"title":"A segmentation method for oral CBCT image based on Segment Anything Model and semi-supervised teacher-student model.","authors":"Jianhong Gan, Runqing Kang, Xun Deng, Tongli He, Nie Yu, Yuling Gan, Peiyang Wei, Xiangyi Chen, Xiaoli Peng, Zhibin Li","doi":"10.1002/mp.17854","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Accurate segmentation of oral cone beam computed tomography (CBCT) images is essential for research and clinical diagnosis. However, irregular and blurred tooth boundaries in CBCT images complicate the labeling of oral tissues, and insufficient labeled samples further limit the generalization ability of segmentation models. The Segment Anything Model (SAM) demonstrates strong generalization and segmentation accuracy across diverse tasks as a vision foundation model. The Teacher-Student (TS) model has proven effective in semi-supervised learning approaches.</p><p><strong>Purpose: </strong>To accurately segment various parts of oral CBCT, such as enamel, pulp, bone, blood vessels, air, etc., an improved segmentation method named SAM-TS is proposed, which combines SAM with the TS model. SAM-TS leverages Low-Rank Adaptation (LoRA) to fine-tune the SAM model on oral CBCT images with fewer parameters.</p><p><strong>Methods: </strong>To efficiently utilize numerous unlabeled images for training models, the LoRA strategy is improved to fine-tune the SAM. The fine-tuned SAM and teacher models collaboratively generate pseudo-labels on unlabeled images, which are filtered and utilized to train the student model. Then, a data augmentation-based Mean Intersection over Union (MIoU) method is proposed to filter out unreliable or spurious pseudo-labels. Finally, the Exponential Moving Average (EMA) method is used to transfer the student model's parameters to the teacher model. After repeating this process, the final optimized student model for segmentation is obtained. The experimental results demonstrate that incorporating unlabeled data into model training through SAM-TS significantly enhances the model's generalization ability and segmentation accuracy.</p><p><strong>Results: </strong>Compared to the baseline algorithm, the proposed method achieves an overall improvement of over 6.48% in MIoU. In the tooth segmentation task, the minimum MIoU and maximum MIoU increased by at least 10% and 27.32%, respectively. In the bone segmentation task, the minimum MIoU and maximum MIoU increased by 7.9% and 32.44%, respectively. Additionally, for overall segmentation, the Hausdorff distance (HD) decreased by 5.1 mm, and the Dice coefficient increased by 2.87%.</p><p><strong>Conclusion: </strong>SAM-TS outperforms existing semi-supervised methods, offering a more competitive and efficient approach to CBCT image segmentation. This method addresses the data annotation bottleneck and opens new avenues for semi-supervised learning applications in medical imaging.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/mp.17854","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Accurate segmentation of oral cone beam computed tomography (CBCT) images is essential for research and clinical diagnosis. However, irregular and blurred tooth boundaries in CBCT images complicate the labeling of oral tissues, and insufficient labeled samples further limit the generalization ability of segmentation models. The Segment Anything Model (SAM) demonstrates strong generalization and segmentation accuracy across diverse tasks as a vision foundation model. The Teacher-Student (TS) model has proven effective in semi-supervised learning approaches.

Purpose: To accurately segment various parts of oral CBCT, such as enamel, pulp, bone, blood vessels, air, etc., an improved segmentation method named SAM-TS is proposed, which combines SAM with the TS model. SAM-TS leverages Low-Rank Adaptation (LoRA) to fine-tune the SAM model on oral CBCT images with fewer parameters.

Methods: To efficiently utilize numerous unlabeled images for training models, the LoRA strategy is improved to fine-tune the SAM. The fine-tuned SAM and teacher models collaboratively generate pseudo-labels on unlabeled images, which are filtered and utilized to train the student model. Then, a data augmentation-based Mean Intersection over Union (MIoU) method is proposed to filter out unreliable or spurious pseudo-labels. Finally, the Exponential Moving Average (EMA) method is used to transfer the student model's parameters to the teacher model. After repeating this process, the final optimized student model for segmentation is obtained. The experimental results demonstrate that incorporating unlabeled data into model training through SAM-TS significantly enhances the model's generalization ability and segmentation accuracy.

Results: Compared to the baseline algorithm, the proposed method achieves an overall improvement of over 6.48% in MIoU. In the tooth segmentation task, the minimum MIoU and maximum MIoU increased by at least 10% and 27.32%, respectively. In the bone segmentation task, the minimum MIoU and maximum MIoU increased by 7.9% and 32.44%, respectively. Additionally, for overall segmentation, the Hausdorff distance (HD) decreased by 5.1 mm, and the Dice coefficient increased by 2.87%.

Conclusion: SAM-TS outperforms existing semi-supervised methods, offering a more competitive and efficient approach to CBCT image segmentation. This method addresses the data annotation bottleneck and opens new avenues for semi-supervised learning applications in medical imaging.

基于分段任意模型和半监督师生模型的口腔CBCT图像分割方法。
背景:口腔锥束计算机断层扫描(CBCT)图像的准确分割对研究和临床诊断至关重要。然而,由于CBCT图像中牙齿边界的不规则和模糊,使得口腔组织的标记变得复杂,标记样本的不足进一步限制了分割模型的泛化能力。作为一种视觉基础模型,SAM在不同的任务中表现出很强的泛化和分割准确性。师生(TS)模型在半监督学习方法中已被证明是有效的。目的:为了对口腔CBCT各部分如牙釉质、牙髓、骨骼、血管、空气等进行准确分割,提出了一种将SAM与TS模型相结合的改进分割方法SAM-TS。SAM- ts利用低秩自适应(Low-Rank Adaptation, LoRA)对口腔CBCT图像的SAM模型进行微调,使其参数更少。方法:为了有效地利用大量未标记图像进行训练模型,改进了LoRA策略,对SAM进行微调。经过微调的SAM和教师模型协同在未标记的图像上生成伪标签,这些伪标签被过滤并用于训练学生模型。然后,提出了一种基于数据增强的平均交联(MIoU)方法来过滤不可靠或虚假的伪标签。最后,使用指数移动平均(EMA)方法将学生模型的参数转移到教师模型中。重复此过程后,得到最终优化的分割学生模型。实验结果表明,通过SAM-TS将未标记数据纳入模型训练,可以显著提高模型的泛化能力和分割精度。结果:与基线算法相比,本文提出的方法在MIoU上总体提高了6.48%以上。在牙齿分割任务中,最小MIoU和最大MIoU分别增加了至少10%和27.32%。在骨分割任务中,最小MIoU和最大MIoU分别增加了7.9%和32.44%。整体分割时,Hausdorff距离(HD)减小5.1 mm, Dice系数增大2.87%。结论:SAM-TS方法优于现有的半监督方法,为CBCT图像分割提供了更具竞争力和效率的方法。该方法解决了数据标注的瓶颈,为半监督学习在医学成像中的应用开辟了新的途径。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信