Jian Pan, Ruijuan Lv, Qun Wang, Xiaobin Zhao, Jiangang Liu, Lin Ai
{"title":"Discrimination between leucine-rich glioma-inactivated 1 antibody encephalitis and gamma-aminobutyric acid B receptor antibody encephalitis based on ResNet18.","authors":"Jian Pan, Ruijuan Lv, Qun Wang, Xiaobin Zhao, Jiangang Liu, Lin Ai","doi":"10.1186/s42492-023-00144-5","DOIUrl":"10.1186/s42492-023-00144-5","url":null,"abstract":"<p><p>This study aims to discriminate between leucine-rich glioma-inactivated 1 (LGI1) antibody encephalitis and gamma-aminobutyric acid B (GABAB) receptor antibody encephalitis using a convolutional neural network (CNN) model. A total of 81 patients were recruited for this study. ResNet18, VGG16, and ResNet50 were trained and tested separately using 3828 positron emission tomography image slices that contained the medial temporal lobe (MTL) or basal ganglia (BG). Leave-one-out cross-validation at the patient level was used to evaluate the CNN models. The receiver operating characteristic (ROC) curve and the area under the ROC curve (AUC) were generated to evaluate the CNN models. Based on the prediction results at slice level, a decision strategy was employed to evaluate the CNN models' performance at patient level. The ResNet18 model achieved the best performance at the slice (AUC = 0.86, accuracy = 80.28%) and patient levels (AUC = 0.98, accuracy = 96.30%). Specifically, at the slice level, 73.28% (1445/1972) of image slices with GABAB receptor antibody encephalitis and 87.72% (1628/1856) of image slices with LGI1 antibody encephalitis were accurately detected. At the patient level, 94.12% (16/17) of patients with GABAB receptor antibody encephalitis and 96.88% (62/64) of patients with LGI1 antibody encephalitis were accurately detected. Heatmaps of the image slices extracted using gradient-weighted class activation mapping indicated that the model focused on the MTL and BG for classification. In general, the ResNet18 model is a potential approach for discriminating between LGI1 and GABAB receptor antibody encephalitis. Metabolism in the MTL and BG is important for discriminating between these two encephalitis subtypes.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"17"},"PeriodicalIF":0.0,"publicationDate":"2023-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10435436/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10421807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hyperparameter optimization for cardiovascular disease data-driven prognostic system.","authors":"Jayson Saputra, Cindy Lawrencya, Jecky Mitra Saini, Suharjito Suharjito","doi":"10.1186/s42492-023-00143-6","DOIUrl":"https://doi.org/10.1186/s42492-023-00143-6","url":null,"abstract":"<p><p>Prediction and diagnosis of cardiovascular diseases (CVDs) based, among other things, on medical examinations and patient symptoms are the biggest challenges in medicine. About 17.9 million people die from CVDs annually, accounting for 31% of all deaths worldwide. With a timely prognosis and thorough consideration of the patient's medical history and lifestyle, it is possible to predict CVDs and take preventive measures to eliminate or control this life-threatening disease. In this study, we used various patient datasets from a major hospital in the United States as prognostic factors for CVD. The data was obtained by monitoring a total of 918 patients whose criteria for adults were 28-77 years old. In this study, we present a data mining modeling approach to analyze the performance, classification accuracy and number of clusters on Cardiovascular Disease Prognostic datasets in unsupervised machine learning (ML) using the Orange data mining software. Various techniques are then used to classify the model parameters, such as k-nearest neighbors, support vector machine, random forest, artificial neural network (ANN), naïve bayes, logistic regression, stochastic gradient descent (SGD), and AdaBoost. To determine the number of clusters, various unsupervised ML clustering methods were used, such as k-means, hierarchical, and density-based spatial clustering of applications with noise clustering. The results showed that the best model performance analysis and classification accuracy were SGD and ANN, both of which had a high score of 0.900 on Cardiovascular Disease Prognostic datasets. Based on the results of most clustering methods, such as k-means and hierarchical clustering, Cardiovascular Disease Prognostic datasets can be divided into two clusters. The prognostic accuracy of CVD depends on the accuracy of the proposed model in determining the diagnostic model. The more accurate the model, the better it can predict which patients are at risk for CVD.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"16"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10390457/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9925692","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Survey of methods and principles in three-dimensional reconstruction from two-dimensional medical images.","authors":"Mriganka Sarmah, Arambam Neelima, Heisnam Rohen Singh","doi":"10.1186/s42492-023-00142-7","DOIUrl":"https://doi.org/10.1186/s42492-023-00142-7","url":null,"abstract":"<p><p>Three-dimensional (3D) reconstruction of human organs has gained attention in recent years due to advances in the Internet and graphics processing units. In the coming years, most patient care will shift toward this new paradigm. However, development of fast and accurate 3D models from medical images or a set of medical scans remains a daunting task due to the number of pre-processing steps involved, most of which are dependent on human expertise. In this review, a survey of pre-processing steps was conducted, and reconstruction techniques for several organs in medical diagnosis were studied. Various methods and principles related to 3D reconstruction were highlighted. The usefulness of 3D reconstruction of organs in medical diagnosis was also highlighted.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"15"},"PeriodicalIF":0.0,"publicationDate":"2023-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10371974/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9886768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Vision transformer architecture and applications in digital health: a tutorial and survey.","authors":"Khalid Al-Hammuri, Fayez Gebali, Awos Kanan, Ilamparithi Thirumarai Chelvan","doi":"10.1186/s42492-023-00140-9","DOIUrl":"10.1186/s42492-023-00140-9","url":null,"abstract":"<p><p>The vision transformer (ViT) is a state-of-the-art architecture for image recognition tasks that plays an important role in digital health applications. Medical images account for 90% of the data in digital medicine applications. This article discusses the core foundations of the ViT architecture and its digital health applications. These applications include image segmentation, classification, detection, prediction, reconstruction, synthesis, and telehealth such as report generation and security. This article also presents a roadmap for implementing the ViT in digital health systems and discusses its limitations and challenges.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"14"},"PeriodicalIF":0.0,"publicationDate":"2023-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10333157/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9809711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"DB-DCAFN: dual-branch deformable cross-attention fusion network for bacterial segmentation.","authors":"Jingkun Wang, Xinyu Ma, Long Cao, Yilin Leng, Zeyi Li, Zihan Cheng, Yuzhu Cao, Xiaoping Huang, Jian Zheng","doi":"10.1186/s42492-023-00141-8","DOIUrl":"https://doi.org/10.1186/s42492-023-00141-8","url":null,"abstract":"<p><p>Sputum smear tests are critical for the diagnosis of respiratory diseases. Automatic segmentation of bacteria from sputum smear images is important for improving diagnostic efficiency. However, this remains a challenging task owing to the high interclass similarity among different categories of bacteria and the low contrast of the bacterial edges. To explore more levels of global pattern features to promote the distinguishing ability of bacterial categories and maintain sufficient local fine-grained features to ensure accurate localization of ambiguous bacteria simultaneously, we propose a novel dual-branch deformable cross-attention fusion network (DB-DCAFN) for accurate bacterial segmentation. Specifically, we first designed a dual-branch encoder consisting of multiple convolution and transformer blocks in parallel to simultaneously extract multilevel local and global features. We then designed a sparse and deformable cross-attention module to capture the semantic dependencies between local and global features, which can bridge the semantic gap and fuse features effectively. Furthermore, we designed a feature assignment fusion module to enhance meaningful features using an adaptive feature weighting strategy to obtain more accurate segmentation. We conducted extensive experiments to evaluate the effectiveness of DB-DCAFN on a clinical dataset comprising three bacterial categories: Acinetobacter baumannii, Klebsiella pneumoniae, and Pseudomonas aeruginosa. The experimental results demonstrate that the proposed DB-DCAFN outperforms other state-of-the-art methods and is effective at segmenting bacteria from sputum smear images.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"13"},"PeriodicalIF":0.0,"publicationDate":"2023-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10319695/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10117038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Editorial: advances in deep learning techniques for biomedical imaging.","authors":"Chuang Niu, Ge Wang","doi":"10.1186/s42492-023-00139-2","DOIUrl":"https://doi.org/10.1186/s42492-023-00139-2","url":null,"abstract":"","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"12"},"PeriodicalIF":0.0,"publicationDate":"2023-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10284748/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9700994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Beyond the horizon: immersive developments for animal ecology research.","authors":"Ying Zhang, Karsten Klein, Falk Schreiber, Kamran Safi","doi":"10.1186/s42492-023-00138-3","DOIUrl":"https://doi.org/10.1186/s42492-023-00138-3","url":null,"abstract":"<p><p>More diverse data on animal ecology are now available. This \"data deluge\" presents challenges for both biologists and computer scientists; however, it also creates opportunities to improve analysis and answer more holistic research questions. We aim to increase awareness of the current opportunity for interdisciplinary research between animal ecology researchers and computer scientists. Immersive analytics (IA) is an emerging research field in which investigations are performed into how immersive technologies, such as large display walls and virtual reality and augmented reality devices, can be used to improve data analysis, outcomes, and communication. These investigations have the potential to reduce the analysis effort and widen the range of questions that can be addressed. We propose that biologists and computer scientists combine their efforts to lay the foundation for IA in animal ecology research. We discuss the potential and the challenges and outline a path toward a structured approach. We imagine that a joint effort would combine the strengths and expertise of both communities, leading to a well-defined research agenda and design space, practical guidelines, robust and reusable software frameworks, reduced analysis effort, and better comparability of results.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"11"},"PeriodicalIF":0.0,"publicationDate":"2023-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10281911/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10066285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Systematic review of digital twin technology and applications.","authors":"Jun-Feng Yao, Yong Yang, Xue-Cheng Wang, Xiao-Peng Zhang","doi":"10.1186/s42492-023-00137-4","DOIUrl":"https://doi.org/10.1186/s42492-023-00137-4","url":null,"abstract":"<p><p>As one of the most important applications of digitalization, intelligence, and service, the digital twin (DT) breaks through the constraints of time, space, cost, and security on physical entities, expands and optimizes the relevant functions of physical entities, and enhances their application value. This phenomenon has been widely studied in academia and industry. In this study, the concept and definition of DT, as utilized by scholars and researchers in various fields of industry, are summarized. The internal association between DT and related technologies is explained. The four stages of DT development history are identified. The fundamentals of the technology, evaluation indexes, and model frameworks are reviewed. Subsequently, a conceptual ternary model of DT based on time, space, and logic is proposed. The technology and application status of typical DT systems are described. Finally, the current technical challenges of DT technology are analyzed, and directions for future development are discussed.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"10"},"PeriodicalIF":0.0,"publicationDate":"2023-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10229487/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9559194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Qing Lyu, Josh Tan, Michael E Zapadka, Janardhana Ponnatapura, Chuang Niu, Kyle J Myers, Ge Wang, Christopher T Whitlow
{"title":"Translating radiology reports into plain language using ChatGPT and GPT-4 with prompt learning: results, limitations, and potential.","authors":"Qing Lyu, Josh Tan, Michael E Zapadka, Janardhana Ponnatapura, Chuang Niu, Kyle J Myers, Ge Wang, Christopher T Whitlow","doi":"10.1186/s42492-023-00136-5","DOIUrl":"10.1186/s42492-023-00136-5","url":null,"abstract":"<p><p>The large language model called ChatGPT has drawn extensively attention because of its human-like expression and reasoning abilities. In this study, we investigate the feasibility of using ChatGPT in experiments on translating radiology reports into plain language for patients and healthcare providers so that they are educated for improved healthcare. Radiology reports from 62 low-dose chest computed tomography lung cancer screening scans and 76 brain magnetic resonance imaging metastases screening scans were collected in the first half of February for this study. According to the evaluation by radiologists, ChatGPT can successfully translate radiology reports into plain language with an average score of 4.27 in the five-point system with 0.08 places of information missing and 0.07 places of misinformation. In terms of the suggestions provided by ChatGPT, they are generally relevant such as keeping following-up with doctors and closely monitoring any symptoms, and for about 37% of 138 cases in total ChatGPT offers specific suggestions based on findings in the report. ChatGPT also presents some randomness in its responses with occasionally over-simplified or neglected information, which can be mitigated using a more detailed prompt. Furthermore, ChatGPT results are compared with a newly released large model GPT-4, showing that GPT-4 can significantly improve the quality of translated reports. Our results show that it is feasible to utilize large language models in clinical education, and further efforts are needed to address limitations and maximize their potential.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"9"},"PeriodicalIF":0.0,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10192466/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9491321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"EM-Gaze: eye context correlation and metric learning for gaze estimation.","authors":"Jinchao Zhou, Guoan Li, Feng Shi, Xiaoyan Guo, Pengfei Wan, Miao Wang","doi":"10.1186/s42492-023-00135-6","DOIUrl":"10.1186/s42492-023-00135-6","url":null,"abstract":"<p><p>In recent years, deep learning techniques have been used to estimate gaze-a significant task in computer vision and human-computer interaction. Previous studies have made significant achievements in predicting 2D or 3D gazes from monocular face images. This study presents a deep neural network for 2D gaze estimation on mobile devices. It achieves state-of-the-art 2D gaze point regression error, while significantly improving gaze classification error on quadrant divisions of the display. To this end, an efficient attention-based module that correlates and fuses the left and right eye contextual features is first proposed to improve gaze point regression performance. Subsequently, through a unified perspective for gaze estimation, metric learning for gaze classification on quadrant divisions is incorporated as additional supervision. Consequently, both gaze point regression and quadrant classification performances are improved. The experiments demonstrate that the proposed method outperforms existing gaze-estimation methods on the GazeCapture and MPIIFaceGaze datasets.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"8"},"PeriodicalIF":0.0,"publicationDate":"2023-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10163188/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9430292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}