Babak Afshin-Pour , Michael Qiu , Shahrzad Hosseini Vajargah , Helen Cheyne , Kevin Ha , Molly Stewart , Jan Horsky , Rachel Aviv , Nasen Zhang , Mangala Narasimhan , John Chelico , Gabriel Musso , Negin Hajizadeh
{"title":"Discriminating Acute Respiratory Distress Syndrome from other forms of respiratory failure via iterative machine learning","authors":"Babak Afshin-Pour , Michael Qiu , Shahrzad Hosseini Vajargah , Helen Cheyne , Kevin Ha , Molly Stewart , Jan Horsky , Rachel Aviv , Nasen Zhang , Mangala Narasimhan , John Chelico , Gabriel Musso , Negin Hajizadeh","doi":"10.1016/j.ibmed.2023.100087","DOIUrl":"10.1016/j.ibmed.2023.100087","url":null,"abstract":"<div><p>Acute Respiratory Distress Syndrome (ARDS) is associated with high morbidity and mortality. Identification of ARDS enables lung protective strategies, quality improvement interventions, and clinical trial enrolment, but remains challenging particularly in the first 24 hours of mechanical ventilation. To address this we built an algorithm capable of discriminating ARDS from other similarly presenting disorders immediately following mechanical ventilation. Specifically, a clinical team examined medical records from 1263 ICU-admitted, mechanically ventilated patients, retrospectively assigning each patient a diagnosis of “ARDS” or “non-ARDS” (e.g., pulmonary edema). Exploiting data readily available in the clinical setting, including patient demographics, laboratory test results from before the initiation of mechanical ventilation, and features extracted by natural language processing of radiology reports, we applied an iterative pre-processing and machine learning framework. The resulting model successfully discriminated ARDS from non-ARDS causes of respiratory failure (AUC = 0.85) among patients meeting Berlin criteria for severe hypoxia. This analysis also highlighted novel patient variables that were informative for identifying ARDS in ICU settings.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"7 ","pages":"Article 100087"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9812471/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10665721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Machine learning algorithms for classifying corneas by Zernike descriptors","authors":"María S. del Río , Juan P. Trevino","doi":"10.1016/j.ibmed.2022.100081","DOIUrl":"https://doi.org/10.1016/j.ibmed.2022.100081","url":null,"abstract":"<div><p>Keratoconus is the most common primary ectasia, as the treatment is not easy, its early diagnosis is essential. The main goal of this study is to develop a method for classification of specific types of corneal shapes where 55 Zernike coefficients (angular index <em>m</em> = 9) are used as inputs. We describe and apply six Machine Learning (ML) classification methods and an ensemble of them to objectively discriminate between keratoconic and non-keratoconic corneal shapes. Earlier attempts by other authors have successfully implemented several Machine Learning models using different parameters (usually, indirect measurements) and have obtained positive results. Given the importance and ubiquity of Zernike polynomials in the eye care community, our proposal should be a suitable choice to incorporate to current methods which might serve as a prescreening test. In this project we work with 475 corneas, classified by experts in two groups, 50 keratoconics and 425 non-keratoconics. All six models yield high rated results with accuracies above 98%, precisions above 97%, or sensitivities above 93%. Also, by building an assembly with the models, we further improve the accuracy of our classification, for example we found an accuracy of 99.7%, a precision of 99.8% and sensitivity of 98.3%. The model can be easily implemented in any system, being very simple to use, thus providing ophthalmologists with a effortless and powerful tool to make a first diagnosis.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"7 ","pages":"Article 100081"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49857634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Adrito Das , Danyal Z. Khan , John G. Hanrahan , Hani J. Marcus , Danail Stoyanov
{"title":"Automatic generation of operation notes in endoscopic pituitary surgery videos using workflow recognition","authors":"Adrito Das , Danyal Z. Khan , John G. Hanrahan , Hani J. Marcus , Danail Stoyanov","doi":"10.1016/j.ibmed.2023.100107","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100107","url":null,"abstract":"<div><p>Operation notes are a crucial component of patient care. However, writing them manually is prone to human error, particularly in high pressured clinical environments. Automatic generation of operation notes from video recordings can alleviate some of the administrative burdens, improve accuracy, and provide additional information. To achieve this for endoscopic pituitary surgery, 27-steps were identified via expert consensus. Then, for the 97-videos recorded for this study, a timestamp of each step was annotated by an expert surgeon. To automatically determine whether a step is present in a video, a three-stage architecture was created. Firstly, for each step, a convolution neural network was used for binary image classification on each frame of a video. Secondly, for each step, the binary frame classifications were passed to a discriminator for binary video classification. Thirdly, for each video, the binary video classifications were passed to an accumulator for multi-label step classification. The architecture was trained on 77-videos, and tested on 20-videos, where a 0.80 weighted-<em>F</em><sub>1</sub> score was achieved. The classifications were inputted into a clinically based predefined template, and further enriched with additional video analytics. This work therefore demonstrates automatic generation of operative notes from surgical videos is feasible, and can assist surgeons during documentation.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100107"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shahrzad Moinian , Nyoman D. Kurniawan , Shekhar S. Chandra , Viktor Vegh , David C. Reutens
{"title":"An unsupervised deep learning-based image translation method for retrospective motion correction of high resolution kidney MRI","authors":"Shahrzad Moinian , Nyoman D. Kurniawan , Shekhar S. Chandra , Viktor Vegh , David C. Reutens","doi":"10.1016/j.ibmed.2023.100108","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100108","url":null,"abstract":"<div><p>A primary challenge for <em>in vivo</em> kidney magnetic resonance imaging (MRI) is the presence of different types of involuntary physiological motion, affecting the diagnostic utility of acquired images due to severe motion artifacts. Existing prospective and retrospective motion correction methods remain ineffective when dealing with complex large amplitude nonrigid motion artifacts. Here, we introduce an unsupervised deep learning-based image to image translation method between motion-affected and motion-free image domains, for correction of rigid-body, respiratory and nonrigid motion artifacts in <em>vivo</em> kidney MRI.</p><p>High resolution (i.e., 156 × 156 × 370 μm) <em>ex vivo</em> 3 Tesla MRI scans of 13 porcine kidneys (because of their anatomical homology to human kidney) were conducted using a 3D T2-weighted turbo spin echo sequence. Rigid-body, respiratory and nonrigid motion-affected images were then simulated using the magnitude-only <em>ex vivo</em> motion-free image set. Each 2D coronal slice of motion-affected and motion-free image volume was then divided into patches of 128 × 128 for training the model. We proposed to add normalised cross-correlation loss to cycle consistency generative adversarial network structure (NCC-CycleGAN), to enforce edge alignment between motion-corrected and motion-free image domains.</p><p>Our NCC-CycleGAN motion correction model demonstrated high performance with an in-tissue structural similarity index measure of 0.77 ± 0.08, peak signal-to-noise ratio of 26.67 ± 3.44 and learned perceptual image patch similarity of 0.173 ± 0.05 between the reconstructed motion-corrected and ground truth motion-free images. This corresponds to a significant respective average improvement of 34%, 23% and 39% (p < 0.05; paired <em>t</em>-test) for the three metrics to correct the three different types of simulated motion artifacts.</p><p>We demonstrated the feasibility of developing an unsupervised deep learning-based method for efficient automated retrospective kidney MRI motion correction, while preserving microscopic tissue structures in high resolution imaging.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100108"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sex, ethnicity, and race data are often unreported in artificial intelligence and machine learning studies in medicine","authors":"Mahmoud Elmahdy, Ronnie Sebro","doi":"10.1016/j.ibmed.2023.100113","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100113","url":null,"abstract":"<div><p>The use of artificial intelligence (AI) programs in healthcare and medicine has steadily increased over the past decade. One major challenge affecting the use of AI programs is that the results of AI programs are sometimes not replicable, meaning that the performance of the AI program is substantially different in the external testing dataset when compared to its performance in the training or validation datasets. This often happens when the external testing dataset is very different from the training or validation datasets. Sex, ethnicity, and race are some of the most important biological and social determinants of health, and are important factors that may differ between training, validation, and external testing datasets, and may contribute to the lack of reproducibility of AI programs. We reviewed over 28,000 original research articles published in the three journals with the highest impact factors in each of 16 medical specialties between 2019 and 2022, to evaluate how often the sex, ethnic, and racial compositions of the datasets used to develop AI algorithms were reported. We also reviewed all currently used AI reporting guidelines, to evaluate which guidelines recommend specific reporting of sex, ethnicity, and race. We find that only 42.47 % (338/797) of articles reported sex, 1.4 % (12/831) reported ethnicity, and 7.3 % (61/831) reported race. When sex was reported, approximately 55.8 % of the study participants were female, and when ethnicity was reported, only 6.2 % of the study participants were Hispanic/Latino. When race was reported, only 29.4 % of study participants were non-White. Most AI guidelines (93.3 %; 14/15) also did not recommend reporting sex, ethnicity, and race. To have fair and ethnical AI, it is important that the sex, ethnic, and racial compositions of the datasets used to develop the AI program are known.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100113"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000273/pdfft?md5=070012db350ebd8eac9219c40819eaa8&pid=1-s2.0-S2666521223000273-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92045506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Early detection of neurological abnormalities using a combined phase space reconstruction and deep learning approach","authors":"Amjed Al Fahoum, Ala’a Zyout","doi":"10.1016/j.ibmed.2023.100123","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100123","url":null,"abstract":"<div><p>The scientific literature on depression detection using electroencephalogram (EEG) signals is extensive and offers numerous innovative approaches. However, these existing state-of-the-art (SOTA) have limitations that hinder their overall efficacy. They rely significantly on datasets with limited scope and accessibility, which introduces potential biases and diminishes generalizability. In addition, they concentrate on analyzing a single dataset, potentially overlooking the inherent variability and complexity of EEG patterns associated with depression. Moreover, certain SOTA methods employ deep learning architectures with exponential time complexity, resulting in computationally intensive and time-consuming training procedures. Therefore, their practicability and applicability in real-world scenarios are compromised. To address these limitations, a novel integrated methodology that combines the advantages of phase space reconstruction and deep neural networks is proposed. It employs publicly available EEG datasets, mitigating the inherent biases of exclusive data sources. Moreover, the method incorporates reconstructed phase space analysis, a feature engineering technique that captures more accurately the complex EEG patterns associated with depression. Simultaneously, the incorporation of a deep neural network component guarantees optimal efficiency and accurate, seamless classification. Using publicly available datasets, cross-dataset validation, and a novel combination of reconstructed phase space analysis and deep neural networks, the proposed method circumvents the shortcomings of current state-of-the-art (SOTA) approaches. This innovation represents a significant advance in enhancing the accuracy of depression detection and provides the base for EEG-based depression assessment tools applicable to real-world settings. The findings of the study provide a more robust and efficient model, which increases classification precision and decreases computing burden. The study findings layout the foundation for scalable, accessible mental health solutions, identification of the pathological deficits in affected brain tissues, and demonstrate the potential of technology-driven approaches to support and guide depressed individuals and enhance mental health outcomes.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100123"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000376/pdfft?md5=49e91661bcf14318d233d3bb140064a2&pid=1-s2.0-S2666521223000376-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138466610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yoon-Seop Kim , Min Woong Kim , Je Seop Lee , Hee Seung Kang , Erdenebayar Urtnasan , Jung Woo Lee , Ji Hun Kim
{"title":"Development of an artificial intelligence model for triage in a military emergency department: Focusing on abdominal pain in soldiers","authors":"Yoon-Seop Kim , Min Woong Kim , Je Seop Lee , Hee Seung Kang , Erdenebayar Urtnasan , Jung Woo Lee , Ji Hun Kim","doi":"10.1016/j.ibmed.2023.100112","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100112","url":null,"abstract":"<div><h3>Background</h3><p>In military settings, determining whether a patient with abdominal pain requires emergency care can be challenging due to the absence or inexperience of medical staff. Misjudging the severity of abdominal pain can lead to delayed treatment or unnecessary transfers, both of which consume valuable resources. Therefore, our aim was to develop an artificial intelligence model capable of classifying the urgency of abdominal pain cases, taking into account patient characteristics.</p></div><div><h3>Methods</h3><p>We collected structured and unstructured data from patients with abdominal pain visiting South Korean military hospital emergency rooms between January 2015 and 2020. After excluding patients with missing values, 20,432 patients were enrolled. Structured data consisted of age, sex, vital signs, past medical history, and symptoms, while unstructured data included preprocessed free text descriptions of chief complaints and present illness. Patients were divided into training, validation, and test datasets in an 8:1:1 ratio. Using structured data, we developed four conventional machine learning models and a novel mixed model, which combined one of the best performing machine learning models with emergency medical knowledge. And we also created a deep learning model using both structured and unstructured data.</p></div><div><h3>Results</h3><p>Xgboost demonstrated the highest performance among the six models, <u>with an area under the precision-recall curve (AUPRC) score of 0.61. The other five models achieved AUPRC scores as follows: logistic regression (0.24), decision tree (0.22), multi-layer perceptron (0.21), deep neural network (0.58), and mixed model (0.58).</u></p></div><div><h3>Conclusion</h3><p>This study is the first to develop an AI model for identifying emergency cases of abdominal pain in a military setting. With more balanced and better-structured datasets, clinically significant AI model could be developed based on the findings of this study.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100112"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ean S. Bett , Timothy C. Frommeyer , Tejaswini Reddy , James “Ty” Johnson
{"title":"Assessment of patient perceptions of technology and the use of machine-based learning in a clinical encounter","authors":"Ean S. Bett , Timothy C. Frommeyer , Tejaswini Reddy , James “Ty” Johnson","doi":"10.1016/j.ibmed.2023.100096","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100096","url":null,"abstract":"<div><h3>Background</h3><p>Electronic health records (EHR) were implemented to improve patient care, reduce healthcare disparities, engage patients and families, improve care coordination, and maintain privacy and security. Unfortunately, the mandated use of EHR has also resulted in significantly increased clerical and administrative burden, with physicians spending an estimated three-fourths of their daily time interacting with the EHR, which negatively affects within-clinic processes and contributes to burnout. In-room scribes are associated with improvement in all aspects of physician satisfaction and increased productivity, though less is known about the use of other technologies such as Google Glass (GG), Natural Language Processing (NLP) and Machine-Based Learning (MBL) systems. Given the need to decrease administrative burden on clinicians, particularly in the utilization of the EHR, there is a need to explore the intersection between varying degrees of technology in the clinical encounter and their ability to meet the aforementioned goals of the EHR.</p></div><div><h3>Aims</h3><p>The primary aim is to determine predictors of overall perception of care dependent on varying mechanisms used for documentation and medical decision-making in a routine clinical encounter. Secondary aims include comparing the perception of individual vignettes based on demographics of the participants and investigating any differences in perception questions by demographics of the participants.</p></div><div><h3>Methods</h3><p>Video vignettes were shown to 498 OhioHealth Physician Group patients and to ResearchMatch volunteers during a 15-month period following IRB approval. Data included a baseline survey to gather demographic and background characteristics and then a perceptual survey where patients rated the physician in the video on 5 facets using a 1 to 5 Likert scale. The analysis included summarizing data of all continuous and categorical variables as well as overall perceptions analyzed using multivariate linear regression with perception score as the outcome variable.</p></div><div><h3>Results</h3><p>Univariate modeling identified sex, education, and type of technology as three factors that were statistically significantly related to the overall perception score. Males had higher scores than females (p = 0.03) and those with lower education had higher scores (p < 0.001). In addition, the physician documenting outside of the room encounter had statistically significantly higher overall perception scores (mean = 22.2, p < 0.001) and the physician documenting in the room encounter had statistically significantly lower overall perception scores (mean = 15.3, p < 0.001) when compared to the other vignettes. Multivariable modeling identified all three of the univariably significant factors as independent factors related to overall perception score. Specifically, high school education had higher scores than associate/bachelor education (LSM = 21.6 vs. ","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"7 ","pages":"Article 100096"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49857365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Physician leadership in the new era of AI and digital health tools","authors":"Jesse Ehrenfeld","doi":"10.1016/j.ibmed.2023.100109","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100109","url":null,"abstract":"<div><p>The American Medical Association's latest survey on digital health trends showed that adoption of digital tools has grown significantly in the past 3–4 yrs, among all physicians, regardless of gender, specialty or age. This article examines digital health trends, including AI, and the AMA's role in ensuring that physicians are actively involved in the creation of new technologies and innovations in medicine. Physicians understand the potential for new digital tools to address health disparities for patients and streamline our workflow better than anyone.</p></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100109"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49869239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Luigi De Angelis , Francesco Baglivo , Guglielmo Arzilli , Leonardo Calamita , Paolo Ferragina , Gaetano Pierpaolo Privitera , Caterina Rizzo
{"title":"Hospital-acquired infections surveillance and prevention: using Natural Language Processing to analyze unstructured text of hospital discharge letters for surgical site infections identification and risk-stratification.","authors":"Luigi De Angelis , Francesco Baglivo , Guglielmo Arzilli , Leonardo Calamita , Paolo Ferragina , Gaetano Pierpaolo Privitera , Caterina Rizzo","doi":"10.1016/j.ibmed.2023.100120","DOIUrl":"https://doi.org/10.1016/j.ibmed.2023.100120","url":null,"abstract":"","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"8 ","pages":"Article 100120"},"PeriodicalIF":0.0,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666521223000340/pdfft?md5=4d1203c98598473fd6b2a278e08e5243&pid=1-s2.0-S2666521223000340-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138558703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}