Pingkun Yan, Ge Wang, Hanqing Chao, Mannudeep K. Kalra
{"title":"Multimodal radiology AI","authors":"Pingkun Yan, Ge Wang, Hanqing Chao, Mannudeep K. Kalra","doi":"10.1016/j.metrad.2023.100019","DOIUrl":null,"url":null,"abstract":"<div><p>The growing armamentarium of artificial intelligence (AI) tools cleared by the United States Food and Drug Administration mostly target a narrow, single imaging modality or data source of information. While imaging technologies continue evolving rapidly, it is recognized that multimodal data provides synergistic information and enables better performance than what is achievable when these modalities are used separately. Deep learning approaches can integrate multimodal data, including not only imaging but also non-imaging modalities such as electronic medical records (EMRs) and genetic profiles. Such convergence advances clinical applications and research for improved effectiveness, especially the prediction of disease risks. This new avenue could address concerns over justification of imaging scans, clinical context-based interpretation of examinations, effectiveness of single modal and multimodal data to influence clinical decision making, as well as prediction of personalized disease risk. In this new era of radiology AI, the paradigm is being shifted from imaging alone AI analytics to multimodal artificial general intelligence (AGI). The heterogeneity of the data and the non-intuitive nature of certain modalities pose major challenges for developing multimodal large AI models and at the same time bring enormous opportunities.</p></div>","PeriodicalId":100921,"journal":{"name":"Meta-Radiology","volume":"1 2","pages":"Article 100019"},"PeriodicalIF":0.0000,"publicationDate":"2023-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Meta-Radiology","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S295016282300019X","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The growing armamentarium of artificial intelligence (AI) tools cleared by the United States Food and Drug Administration mostly target a narrow, single imaging modality or data source of information. While imaging technologies continue evolving rapidly, it is recognized that multimodal data provides synergistic information and enables better performance than what is achievable when these modalities are used separately. Deep learning approaches can integrate multimodal data, including not only imaging but also non-imaging modalities such as electronic medical records (EMRs) and genetic profiles. Such convergence advances clinical applications and research for improved effectiveness, especially the prediction of disease risks. This new avenue could address concerns over justification of imaging scans, clinical context-based interpretation of examinations, effectiveness of single modal and multimodal data to influence clinical decision making, as well as prediction of personalized disease risk. In this new era of radiology AI, the paradigm is being shifted from imaging alone AI analytics to multimodal artificial general intelligence (AGI). The heterogeneity of the data and the non-intuitive nature of certain modalities pose major challenges for developing multimodal large AI models and at the same time bring enormous opportunities.