Fine-tuning large language models for effective nutrition support in residential aged care: a domain expertise approach

Mohammad Alkhalaf, Chao Deng, Jun Shen, Hui-Chen (Rita) Chang, Ping Yu
{"title":"Fine-tuning large language models for effective nutrition support in residential aged care: a domain expertise approach","authors":"Mohammad Alkhalaf, Chao Deng, Jun Shen, Hui-Chen (Rita) Chang, Ping Yu","doi":"10.1101/2024.07.21.24310775","DOIUrl":null,"url":null,"abstract":"Purpose: Malnutrition is a serious health concern, particularly among the older people living in residential aged care facilities. An automated and efficient method is required to identify the individuals afflicted with malnutrition in this setting. The recent advancements in transformer-based large language models (LLMs) equipped with sophisticated context-aware embeddings, such as RoBERTa, have significantly improved machine learning performance, particularly in predictive modelling. Enhancing the embeddings of these models on domain-specific corpora, such as clinical notes, is essential for elevating their performance in clinical tasks. Therefore, our study introduces a novel approach that trains a foundational RoBERTa model on nursing progress notes to develop a RAC domain-specific LLM. The model is further fine-tuned on nursing progress notes to enhance malnutrition identification and prediction in residential aged care setting.\nMethods: We develop our domain-specific model by training the RoBERTa LLM on 500,000 nursing progress notes from residential aged care electronic health records (EHRs). The model embeddings were used for two downstream tasks: malnutrition note identification and malnutrition prediction. Its performance was compared against baseline RoBERTa and BioClinicalBERT. Furthermore, we truncated long sequence text to fit into RoBERTa 512-token sequence length limitation, enabling our model to handle sequences up to1536 tokens.\nResults: Utilizing 5-fold cross-validation for both tasks, our RAC domain-specific LLM demonstrated significantly better performance over other models. In malnutrition note identification, it achieved a slightly higher F1-score of 0.966 compared to other LLMs. In prediction, it achieved significantly higher F1-score of 0.655. We enhanced our model predictive capability by integrating the risk factors extracted from each client notes, creating a combined data layer of structured risk factors and free-text notes. This integration improved the prediction performance, evidenced by an increased F1-score of 0.687.\nConclusion: Our findings suggest that further fine-tuning a large language model on a domain-specific clinical corpus can improve the foundational model performance in clinical tasks. This specialized adaptation significantly improves our domain-specific model performance in tasks such as malnutrition risk identification and malnutrition prediction, making it useful for identifying and predicting malnutrition among older people living in residential aged care or long-term care facilities.","PeriodicalId":501454,"journal":{"name":"medRxiv - Health Informatics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"medRxiv - Health Informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.07.21.24310775","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: Malnutrition is a serious health concern, particularly among the older people living in residential aged care facilities. An automated and efficient method is required to identify the individuals afflicted with malnutrition in this setting. The recent advancements in transformer-based large language models (LLMs) equipped with sophisticated context-aware embeddings, such as RoBERTa, have significantly improved machine learning performance, particularly in predictive modelling. Enhancing the embeddings of these models on domain-specific corpora, such as clinical notes, is essential for elevating their performance in clinical tasks. Therefore, our study introduces a novel approach that trains a foundational RoBERTa model on nursing progress notes to develop a RAC domain-specific LLM. The model is further fine-tuned on nursing progress notes to enhance malnutrition identification and prediction in residential aged care setting. Methods: We develop our domain-specific model by training the RoBERTa LLM on 500,000 nursing progress notes from residential aged care electronic health records (EHRs). The model embeddings were used for two downstream tasks: malnutrition note identification and malnutrition prediction. Its performance was compared against baseline RoBERTa and BioClinicalBERT. Furthermore, we truncated long sequence text to fit into RoBERTa 512-token sequence length limitation, enabling our model to handle sequences up to1536 tokens. Results: Utilizing 5-fold cross-validation for both tasks, our RAC domain-specific LLM demonstrated significantly better performance over other models. In malnutrition note identification, it achieved a slightly higher F1-score of 0.966 compared to other LLMs. In prediction, it achieved significantly higher F1-score of 0.655. We enhanced our model predictive capability by integrating the risk factors extracted from each client notes, creating a combined data layer of structured risk factors and free-text notes. This integration improved the prediction performance, evidenced by an increased F1-score of 0.687. Conclusion: Our findings suggest that further fine-tuning a large language model on a domain-specific clinical corpus can improve the foundational model performance in clinical tasks. This specialized adaptation significantly improves our domain-specific model performance in tasks such as malnutrition risk identification and malnutrition prediction, making it useful for identifying and predicting malnutrition among older people living in residential aged care or long-term care facilities.
微调大型语言模型,为养老院护理提供有效的营养支持:领域专长方法
目的:营养不良是一个严重的健康问题,尤其是居住在养老院的老年人。在这种情况下,需要一种自动高效的方法来识别营养不良者。基于转换器的大型语言模型(LLM)配备了复杂的上下文感知嵌入(如 RoBERTa),其最新进展显著提高了机器学习性能,尤其是在预测建模方面。加强这些模型在特定领域语料库(如临床笔记)中的嵌入对于提高它们在临床任务中的性能至关重要。因此,我们的研究引入了一种新方法,即在护理进展笔记上训练基础 RoBERTa 模型,从而开发出针对特定领域的 RAC LLM。该模型在护理进展记录的基础上进行了进一步的微调,以提高住院养老护理环境中营养不良的识别和预测能力:方法:我们通过对来自养老院电子健康记录(EHR)的 500,000 份护理进展记录训练 RoBERTa LLM,从而开发出针对特定领域的模型。模型嵌入用于两个下游任务:营养不良记录识别和营养不良预测。我们将其性能与基线 RoBERTa 和 BioClinicalBERT 进行了比较。此外,我们对长序列文本进行了截断,以适应 RoBERTa 512 个标记的序列长度限制,从而使我们的模型能够处理多达 1536 个标记的序列:通过对这两项任务进行 5 倍交叉验证,我们的 RAC 特定领域 LLM 的性能明显优于其他模型。在营养不良注释识别方面,与其他 LLM 相比,它的 F1 分数略高,为 0.966。在预测方面,它的 F1 分数明显更高,达到 0.655。我们通过整合从每个客户笔记中提取的风险因素,创建了结构化风险因素和自由文本笔记的组合数据层,从而增强了模型的预测能力。这种整合提高了预测性能,F1 分数提高到了 0.687:我们的研究结果表明,在特定领域的临床语料库上进一步微调大型语言模型可以提高基础模型在临床任务中的性能。在营养不良风险识别和营养不良预测等任务中,这种专业化的调整大大提高了特定领域模型的性能,使其可用于识别和预测居住在养老院或长期护理机构的老年人的营养不良情况。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信