An Application of 3D Vision Transformers and Explainable AI in Prosthetic Dentistry

Faisal Ahmed Sifat, Md Sahadul Hasan Arian, Saif Ahmed, Taseef Hasan Farook, Nabeel Mohammed, James Dudley
{"title":"An Application of 3D Vision Transformers and Explainable AI in Prosthetic Dentistry","authors":"Faisal Ahmed Sifat,&nbsp;Md Sahadul Hasan Arian,&nbsp;Saif Ahmed,&nbsp;Taseef Hasan Farook,&nbsp;Nabeel Mohammed,&nbsp;James Dudley","doi":"10.1002/ail2.101","DOIUrl":null,"url":null,"abstract":"<p>To create and validate a transformer-based deep neural network architecture for classifying 3D scans of teeth for computer-assisted manufacturing and dental prosthetic rehabilitation surpassing previously reported validation accuracies obtained with convolutional neural networks (CNNs). Voxel-based representation and encoding input data in a high-dimensional space forms of preprocessing were investigated using 34 3D models of teeth obtained from intraoral scanning. Independent CNNs and vision transformers (ViTs), and their combination (CNN and ViT hybrid model) were implemented to classify the 3D scans directly from standard tessellation language (.stl) files and an Explainable AI (ExAI) model was generated to qualitatively explore the deterministic patterns that influenced the outcomes of the automation process. The results demonstrate that the CNN and ViT hybrid model architecture surpasses conventional supervised CNN, achieving a consistent validation accuracy of 90% through three-fold cross-validation. This process validated our initial findings, where each instance had the opportunity to be part of the validation set, ensuring it remained unseen during training. Furthermore, employing high-dimensional encoding of input data solely with 3DCNN yields a validation accuracy of 80%. When voxel data preprocessing is utilized, ViT outperforms CNN, achieving validation accuracies of 80% and 50%, respectively. The study also highlighted the saliency map's ability to identify areas of tooth cavity preparation of restorative importance, that can theoretically enable more accurate 3D printed prosthetic outputs. The investigation introduced a CNN and ViT hybrid model for classification of 3D tooth models in digital dentistry, and it was the first to employ ExAI in the efforts to automate the process of dental computer-assisted manufacturing.</p>","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":"5 4","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ail2.101","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied AI letters","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ail2.101","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

To create and validate a transformer-based deep neural network architecture for classifying 3D scans of teeth for computer-assisted manufacturing and dental prosthetic rehabilitation surpassing previously reported validation accuracies obtained with convolutional neural networks (CNNs). Voxel-based representation and encoding input data in a high-dimensional space forms of preprocessing were investigated using 34 3D models of teeth obtained from intraoral scanning. Independent CNNs and vision transformers (ViTs), and their combination (CNN and ViT hybrid model) were implemented to classify the 3D scans directly from standard tessellation language (.stl) files and an Explainable AI (ExAI) model was generated to qualitatively explore the deterministic patterns that influenced the outcomes of the automation process. The results demonstrate that the CNN and ViT hybrid model architecture surpasses conventional supervised CNN, achieving a consistent validation accuracy of 90% through three-fold cross-validation. This process validated our initial findings, where each instance had the opportunity to be part of the validation set, ensuring it remained unseen during training. Furthermore, employing high-dimensional encoding of input data solely with 3DCNN yields a validation accuracy of 80%. When voxel data preprocessing is utilized, ViT outperforms CNN, achieving validation accuracies of 80% and 50%, respectively. The study also highlighted the saliency map's ability to identify areas of tooth cavity preparation of restorative importance, that can theoretically enable more accurate 3D printed prosthetic outputs. The investigation introduced a CNN and ViT hybrid model for classification of 3D tooth models in digital dentistry, and it was the first to employ ExAI in the efforts to automate the process of dental computer-assisted manufacturing.

Abstract Image

三维视觉变形与可解释人工智能在牙科修复中的应用
创建并验证基于变压器的深度神经网络架构,用于对牙齿3D扫描进行分类,用于计算机辅助制造和牙科假肢康复,超过先前报道的卷积神经网络(cnn)获得的验证精度。利用口腔内扫描获得的34个牙齿三维模型,研究了基于体素的高维空间表示和编码输入数据的预处理形式。实现了独立的CNN和视觉转换器(ViT)及其组合(CNN和ViT混合模型),直接从标准细分语言(.stl)文件中对3D扫描进行分类,并生成了可解释的AI (ExAI)模型,以定性地探索影响自动化过程结果的确定性模式。结果表明,CNN和ViT混合模型架构优于传统的有监督CNN,通过三次交叉验证,验证准确率达到90%。这个过程验证了我们最初的发现,其中每个实例都有机会成为验证集的一部分,确保它在训练期间不可见。此外,仅使用3DCNN对输入数据进行高维编码,验证准确率达到80%。当使用体素数据预处理时,ViT优于CNN,验证准确率分别达到80%和50%。该研究还强调了显著性图识别牙腔准备修复重要性区域的能力,理论上可以实现更精确的3D打印假体输出。该研究引入了CNN和ViT混合模型,用于数字牙科中3D牙齿模型的分类,并且首次使用ExAI来实现牙科计算机辅助制造过程的自动化。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信