Clinical Natural Language Processing Workshop最新文献

筛选
英文 中文
Analyzing Text Specific vs Blackbox Fairness Algorithms in Multimodal Clinical NLP 多模态临床NLP中文本特定与黑盒公平性算法的分析
Clinical Natural Language Processing Workshop Pub Date : 2020-11-01 DOI: 10.18653/v1/2020.clinicalnlp-1.33
John Chen, Ian Berlot-Attwell, Safwan Hossain, Xindi Wang, Frank Rudzicz
{"title":"Analyzing Text Specific vs Blackbox Fairness Algorithms in Multimodal Clinical NLP","authors":"John Chen, Ian Berlot-Attwell, Safwan Hossain, Xindi Wang, Frank Rudzicz","doi":"10.18653/v1/2020.clinicalnlp-1.33","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.33","url":null,"abstract":"Clinical machine learning is increasingly multimodal, collected in both structured tabular formats and unstructured forms such as free text. We propose a novel task of exploring fairness on a multimodal clinical dataset, adopting equalized odds for the downstream medical prediction tasks. To this end, we investigate a modality-agnostic fairness algorithm - equalized odds post processing - and compare it to a text-specific fairness algorithm: debiased clinical word embeddings. Despite the fact that debiased word embeddings do not explicitly address equalized odds of protected groups, we show that a text-specific approach to fairness may simultaneously achieve a good balance of performance classical notions of fairness. Our work opens the door for future work at the critical intersection of clinical NLP and fairness.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114429771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
The Chilean Waiting List Corpus: a new resource for clinical Named Entity Recognition in Spanish 智利等候名单语料库:西班牙语临床命名实体识别的新资源
Clinical Natural Language Processing Workshop Pub Date : 2020-11-01 DOI: 10.18653/v1/2020.clinicalnlp-1.32
P. Baez, F. Villena, Matías Rojas, Manuel Durán, J. Dunstan
{"title":"The Chilean Waiting List Corpus: a new resource for clinical Named Entity Recognition in Spanish","authors":"P. Baez, F. Villena, Matías Rojas, Manuel Durán, J. Dunstan","doi":"10.18653/v1/2020.clinicalnlp-1.32","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.32","url":null,"abstract":"In this work we describe the Waiting List Corpus consisting of de-identified referrals for several specialty consultations from the waiting list in Chilean public hospitals. A subset of 900 referrals was manually annotated with 9,029 entities, 385 attributes, and 284 pairs of relations with clinical relevance. A trained medical doctor annotated these referrals, and then together with other three researchers, consolidated each of the annotations. The annotated corpus has nested entities, with 32.2% of entities embedded in other entities. We use this annotated corpus to obtain preliminary results for Named Entity Recognition (NER). The best results were achieved by using a biLSTM-CRF architecture using word embeddings trained over Spanish Wikipedia together with clinical embeddings computed by the group. NER models applied to this corpus can leverage statistics of diseases and pending procedures within this waiting list. This work constitutes the first annotated corpus using clinical narratives from Chile, and one of the few for the Spanish language. The annotated corpus, the clinical word embeddings, and the annotation guidelines are freely released to the research community.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116389448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Utilizing Multimodal Feature Consistency to Detect Adversarial Examples on Clinical Summaries 利用多模态特征一致性检测临床摘要中的对抗性样本
Clinical Natural Language Processing Workshop Pub Date : 2020-11-01 DOI: 10.18653/v1/2020.clinicalnlp-1.29
Wenjie Wang, Youngja Park, Taesung Lee, Ian Molloy, Pengfei Tang, Li Xiong
{"title":"Utilizing Multimodal Feature Consistency to Detect Adversarial Examples on Clinical Summaries","authors":"Wenjie Wang, Youngja Park, Taesung Lee, Ian Molloy, Pengfei Tang, Li Xiong","doi":"10.18653/v1/2020.clinicalnlp-1.29","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.29","url":null,"abstract":"Recent studies have shown that adversarial examples can be generated by applying small perturbations to the inputs such that the well- trained deep learning models will misclassify. With the increasing number of safety and security-sensitive applications of deep learn- ing models, the robustness of deep learning models has become a crucial topic. The robustness of deep learning models for health- care applications is especially critical because the unique characteristics and the high financial interests of the medical domain make it more sensitive to adversarial attacks. Among the modalities of medical data, the clinical summaries have higher risks to be attacked because they are generated by third-party companies. As few works studied adversarial threats on clinical summaries, in this work we first apply adversarial attack to clinical summaries of electronic health records (EHR) to show the text-based deep learning systems are vulnerable to adversarial examples. Secondly, benefiting from the multi-modality of the EHR dataset, we propose a novel defense method, MATCH (Multimodal feATure Consistency cHeck), which leverages the consistency between multiple modalities in the data to defend against adversarial examples on a single modality. Our experiments demonstrate the effectiveness of MATCH on a hospital readmission prediction task comparing with baseline methods.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127462994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Advancing Seq2seq with Joint Paraphrase Learning 用联合释义学习推进Seq2seq
Clinical Natural Language Processing Workshop Pub Date : 2020-11-01 DOI: 10.18653/v1/2020.clinicalnlp-1.30
So Yeon Min, Preethi Raghavan, Peter Szolovits
{"title":"Advancing Seq2seq with Joint Paraphrase Learning","authors":"So Yeon Min, Preethi Raghavan, Peter Szolovits","doi":"10.18653/v1/2020.clinicalnlp-1.30","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.30","url":null,"abstract":"We address the problem of model generalization for sequence to sequence (seq2seq) architectures. We propose going beyond data augmentation via paraphrase-optimized multi-task learning and observe that it is useful in correctly handling unseen sentential paraphrases as inputs. Our models greatly outperform SOTA seq2seq models for semantic parsing on diverse domains (Overnight - up to 3.2% and emrQA - 7%) and Nematus, the winning solution for WMT 2017, for Czech to English translation (CzENG 1.6 - 1.5 BLEU).","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114035045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
PHICON: Improving Generalization of Clinical Text De-identification Models via Data Augmentation 通过数据增强提高临床文本去识别模型的泛化
Clinical Natural Language Processing Workshop Pub Date : 2020-10-11 DOI: 10.18653/v1/2020.clinicalnlp-1.23
Xiang Yue, Shuang Zhou
{"title":"PHICON: Improving Generalization of Clinical Text De-identification Models via Data Augmentation","authors":"Xiang Yue, Shuang Zhou","doi":"10.18653/v1/2020.clinicalnlp-1.23","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.23","url":null,"abstract":"De-identification is the task of identifying protected health information (PHI) in the clinical text. Existing neural de-identification models often fail to generalize to a new dataset. We propose a simple yet effective data augmentation method PHICON to alleviate the generalization issue. PHICON consists of PHI augmentation and Context augmentation, which creates augmented training corpora by replacing PHI entities with named-entities sampled from external sources, and by changing background context with synonym replacement or random word insertion, respectively. Experimental results on the i2b2 2006 and 2014 de-identification challenge datasets show that PHICON can help three selected de-identification models boost F1-score (by at most 8.6%) on cross-dataset test setting. We also discuss how much augmentation to use and how each augmentation method influences the performance.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133506491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Information Extraction from Swedish Medical Prescriptions with Sig-Transformer Encoder 基于符号转换器编码器的瑞典处方信息提取
Clinical Natural Language Processing Workshop Pub Date : 2020-10-10 DOI: 10.18653/v1/2020.clinicalnlp-1.5
John N. Pougue Biyong, Bo Wang, Terry Lyons, A. Nevado-Holgado
{"title":"Information Extraction from Swedish Medical Prescriptions with Sig-Transformer Encoder","authors":"John N. Pougue Biyong, Bo Wang, Terry Lyons, A. Nevado-Holgado","doi":"10.18653/v1/2020.clinicalnlp-1.5","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.5","url":null,"abstract":"Relying on large pretrained language models such as Bidirectional Encoder Representations from Transformers (BERT) for encoding and adding a simple prediction layer has led to impressive performance in many clinical natural language processing (NLP) tasks. In this work, we present a novel extension to the Transformer architecture, by incorporating signature transform with the self-attention model. This architecture is added between embedding and prediction layers. Experiments on a new Swedish prescription data show the proposed architecture to be superior in two of the three information extraction tasks, comparing to baseline models. Finally, we evaluate two different embedding approaches between applying Multilingual BERT and translating the Swedish text to English then encode with a BERT model pretrained on clinical notes.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124969589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Distinguishing between Dementia with Lewy bodies (DLB) and Alzheimer’s Disease (AD) using Mental Health Records: a Classification Approach 用精神健康记录区分路易体痴呆(DLB)和阿尔茨海默病(AD):一种分类方法
Clinical Natural Language Processing Workshop Pub Date : 2020-10-06 DOI: 10.18653/V1/2020.CLINICALNLP-1.19
Zixu Wang, Julia Ive, S. Moylett, C. Mueller, R. Cardinal, S. Velupillai, J. O'Brien, R. Stewart
{"title":"Distinguishing between Dementia with Lewy bodies (DLB) and Alzheimer’s Disease (AD) using Mental Health Records: a Classification Approach","authors":"Zixu Wang, Julia Ive, S. Moylett, C. Mueller, R. Cardinal, S. Velupillai, J. O'Brien, R. Stewart","doi":"10.18653/V1/2020.CLINICALNLP-1.19","DOIUrl":"https://doi.org/10.18653/V1/2020.CLINICALNLP-1.19","url":null,"abstract":"While Dementia with Lewy Bodies (DLB) is the second most common type of neurodegenerative dementia following Alzheimer’s Disease (AD), it is difficult to distinguish from AD. We propose a method for DLB detection by using mental health record (MHR) documents from a (3-month) period before a patient has been diagnosed with DLB or AD. Our objective is to develop a model that could be clinically useful to differentiate between DLB and AD across datasets from different healthcare institutions. We cast this as a classification task using Convolutional Neural Network (CNN), an efficient neural model for text classification. We experiment with different representation models, and explore the features that contribute to model performances. In addition, we apply temperature scaling, a simple but efficient model calibration method, to produce more reliable predictions. We believe the proposed method has important potential for clinical applications using routine healthcare records, and for generalising to other relevant clinical record datasets. To the best of our knowledge, this is the first attempt to distinguish DLB from AD using mental health records, and to improve the reliability of DLB predictions.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132521537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BERT-XML: Large Scale Automated ICD Coding Using BERT Pretraining 使用BERT预训练的大规模自动ICD编码
Clinical Natural Language Processing Workshop Pub Date : 2020-05-26 DOI: 10.18653/v1/2020.clinicalnlp-1.3
Zachariah Zhang, Jingshu Liu, N. Razavian
{"title":"BERT-XML: Large Scale Automated ICD Coding Using BERT Pretraining","authors":"Zachariah Zhang, Jingshu Liu, N. Razavian","doi":"10.18653/v1/2020.clinicalnlp-1.3","DOIUrl":"https://doi.org/10.18653/v1/2020.clinicalnlp-1.3","url":null,"abstract":"ICD coding is the task of classifying and cod-ing all diagnoses, symptoms and proceduresassociated with a patient’s visit. The process isoften manual, extremely time-consuming andexpensive for hospitals as clinical interactionsare usually recorded in free text medical notes.In this paper, we propose a machine learningmodel, BERT-XML, for large scale automatedICD coding of EHR notes, utilizing recentlydeveloped unsupervised pretraining that haveachieved state of the art performance on a va-riety of NLP tasks. We train a BERT modelfrom scratch on EHR notes, learning with vo-cabulary better suited for EHR tasks and thusoutperform off-the-shelf models. We furtheradapt the BERT architecture for ICD codingwith multi-label attention. We demonstratethe effectiveness of BERT-based models on thelarge scale ICD code classification task usingmillions of EHR notes to predict thousands ofunique codes.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132455524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 55
HealthMavericks@MEDIQA-Chat 2023: Benchmarking different Transformer based models for Clinical Dialogue Summarization HealthMavericks@MEDIQA-Chat 2023:基于临床对话总结的不同Transformer模型的基准测试
Clinical Natural Language Processing Workshop Pub Date : 1900-01-01 DOI: 10.18653/v1/2023.clinicalnlp-1.50
Kunal Suri, Saumajit Saha, Ashutosh Kumar Singh
{"title":"HealthMavericks@MEDIQA-Chat 2023: Benchmarking different Transformer based models for Clinical Dialogue Summarization","authors":"Kunal Suri, Saumajit Saha, Ashutosh Kumar Singh","doi":"10.18653/v1/2023.clinicalnlp-1.50","DOIUrl":"https://doi.org/10.18653/v1/2023.clinicalnlp-1.50","url":null,"abstract":"In recent years, we have seen many Transformer based models being created to address Dialog Summarization problem. While there has been a lot of work on understanding how these models stack against each other in summarizing regular conversations such as the ones found in DialogSum dataset, there haven’t been many analysis of these models on Clinical Dialog Summarization. In this article, we describe our solution to MEDIQA-Chat 2023 Shared Tasks as part of ACL-ClinicalNLP 2023 workshop which benchmarks some of the popular Transformer Architectures such as BioBart, Flan-T5, DialogLED, and OpenAI GPT3 on the problem of Clinical Dialog Summarization. We analyse their performance on two tasks - summarizing short conversations and long conversations. In addition to this, we also benchmark two popular summarization ensemble methods and report their performance.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122784459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
clulab at MEDIQA-Chat 2023: Summarization and classification of medical dialogues MEDIQA-Chat 2023:医学对话的总结和分类
Clinical Natural Language Processing Workshop Pub Date : 1900-01-01 DOI: 10.18653/v1/2023.clinicalnlp-1.19
Kadir Bulut Özler, S. Bethard
{"title":"clulab at MEDIQA-Chat 2023: Summarization and classification of medical dialogues","authors":"Kadir Bulut Özler, S. Bethard","doi":"10.18653/v1/2023.clinicalnlp-1.19","DOIUrl":"https://doi.org/10.18653/v1/2023.clinicalnlp-1.19","url":null,"abstract":"Clinical Natural Language Processing has been an increasingly popular research area in the NLP community. With the rise of large language models (LLMs) and their impressive abilities in NLP tasks, it is crucial to pay attention to their clinical applications. Sequence to sequence generative approaches with LLMs have been widely used in recent years. To be a part of the research in clinical NLP with recent advances in the field, we participated in task A of MEDIQA-Chat at ACL-ClinicalNLP Workshop 2023. In this paper, we explain our methods and findings as well as our comments on our results and limitations.","PeriodicalId":216954,"journal":{"name":"Clinical Natural Language Processing Workshop","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121698561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信