Transfer Learning Improves Unsupervised Assignment of ICD codes with Clinical Notes

Amit Kumar, Souparna Das, Suman Roy
{"title":"Transfer Learning Improves Unsupervised Assignment of ICD codes with Clinical Notes","authors":"Amit Kumar, Souparna Das, Suman Roy","doi":"10.1109/ICDH60066.2023.00047","DOIUrl":null,"url":null,"abstract":"In healthcare industry, it is a standard practice to assign a set of International Classification of Diseases (ICD) to a clinical note (which can be a patient visit, a discharge summary and the like) as part of medical coding process mandated by medical care and patient billing. A supervised framework is adopted for most of the automated ICD coding assignment methods in which a subset of the clinical notes are a-priori labeled with ICD codes. But in lot of cases enough labeled texts are not available. These call for an unsupervised assignment of ICD codes. However, the quality of the data plays an important role in the performance of unsupervised coding, - low quality data leads to degradation of performance. In this paper, we explore a transfer learning approach for ICD coding using a combination of pre-training and supervised fine-tuning. We use a hierarchical BERT model comprising of a Bi-LSTM layered on top of BERT (this removes the restriction on the size of clinical texts)) as part of model architecture, and pre-train it on the total corpus (which include both labeled and unlabeled data). Next we transfer its weights to fine tune the model with labeled data (MIMIC data) in a supervised framework and then use this model to predict ICD code for unlabeled data using token similarity. This is the first use of using transfer learning in ICD prediction to our knowledge. Finally we show the efficacy of our transfer learning approach through rigorous experimentation, - there is 20% gain of sensitivity (recall) and 6% lift in specificity in ICD prediction compared to direct unsupervised prediction.","PeriodicalId":107307,"journal":{"name":"2023 IEEE International Conference on Digital Health (ICDH)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Digital Health (ICDH)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDH60066.2023.00047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In healthcare industry, it is a standard practice to assign a set of International Classification of Diseases (ICD) to a clinical note (which can be a patient visit, a discharge summary and the like) as part of medical coding process mandated by medical care and patient billing. A supervised framework is adopted for most of the automated ICD coding assignment methods in which a subset of the clinical notes are a-priori labeled with ICD codes. But in lot of cases enough labeled texts are not available. These call for an unsupervised assignment of ICD codes. However, the quality of the data plays an important role in the performance of unsupervised coding, - low quality data leads to degradation of performance. In this paper, we explore a transfer learning approach for ICD coding using a combination of pre-training and supervised fine-tuning. We use a hierarchical BERT model comprising of a Bi-LSTM layered on top of BERT (this removes the restriction on the size of clinical texts)) as part of model architecture, and pre-train it on the total corpus (which include both labeled and unlabeled data). Next we transfer its weights to fine tune the model with labeled data (MIMIC data) in a supervised framework and then use this model to predict ICD code for unlabeled data using token similarity. This is the first use of using transfer learning in ICD prediction to our knowledge. Finally we show the efficacy of our transfer learning approach through rigorous experimentation, - there is 20% gain of sensitivity (recall) and 6% lift in specificity in ICD prediction compared to direct unsupervised prediction.
迁移学习改进临床记录的ICD代码的无监督分配
在医疗保健行业,将一组国际疾病分类(ICD)分配到临床记录(可以是患者访问、出院摘要等)是一种标准做法,作为医疗保健和患者账单强制要求的医疗编码过程的一部分。大多数自动ICD编码分配方法采用监督框架,其中临床笔记的一个子集先验地标记为ICD代码。但在很多情况下,没有足够的标记文本可用。这些要求对ICD代码进行无监督分配。然而,数据的质量对无监督编码的性能起着重要的作用,低质量的数据会导致性能的下降。在本文中,我们探索了一种使用预训练和监督微调相结合的ICD编码迁移学习方法。我们使用了一个分层的BERT模型,该模型由在BERT之上分层的Bi-LSTM组成(这消除了对临床文本大小的限制),作为模型架构的一部分,并在整个语料库(包括标记和未标记的数据)上对其进行预训练。接下来,我们将其权重转移到有监督框架中带有标记数据(MIMIC数据)的模型中,然后使用该模型使用令牌相似性来预测未标记数据的ICD代码。据我们所知,这是在ICD预测中首次使用迁移学习。最后,我们通过严格的实验证明了我们的迁移学习方法的有效性,与直接无监督预测相比,ICD预测的灵敏度(召回率)提高了20%,特异性提高了6%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信