Using 3D facial information to predict malnutrition and consequent complications.

IF 2.1 4区 医学 Q3 NUTRITION & DIETETICS
Xue Wang, Weijia Wang, Moxi Chen, Meifen Han, Zhiqin Rong, Jin Fu, Yuming Chong, Nanze Yu, Xiao Long, Zhitao Cheng, Yong Tang, Wei Chen
{"title":"Using 3D facial information to predict malnutrition and consequent complications.","authors":"Xue Wang, Weijia Wang, Moxi Chen, Meifen Han, Zhiqin Rong, Jin Fu, Yuming Chong, Nanze Yu, Xiao Long, Zhitao Cheng, Yong Tang, Wei Chen","doi":"10.1002/ncp.11215","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Phase angle (PhA) correlates with body composition and could predict the nutrition status of patients and disease prognosis. We aimed to explore the feasibility of predicting PhA-diagnosed malnutrition using facial image information based on deep learning (DL).</p><p><strong>Methods: </strong>From August 2021 to April 2022, inpatients were enrolled from surgery, gastroenterology, and oncology departments in a tertiary hospital. Subjective global assessment was used as the gold standard of malnutrition diagnosis. The highest Youden index value was selected as the PhA cutoff point. We developed a multimodal DL framework to automatically analyze the three-dimensional (3D) facial data and accurately determine patients' PhA categories. The framework was trained and validated using a cross-validation approach and tested on an independent dataset.</p><p><strong>Results: </strong>Four hundred eighty-two patients were included in the final dataset, including 176 with malnourishment. In male patients, the PhA value with the highest Youden index was 5.55°, and the area under the receiver operating characteristic curve (AUC) = 0.68; in female patients, the PhA value with the highest Youden index was 4.88°, and AUC = 0.69. Inpatients with low PhA had higher incidence of infectious complications during the hospital stay (P = 0.003). The DL model trained with 4096 points extracted from 3D facial data had the best performance. The algorithm showed fair performance in predicting PhA, with an AUC of 0.77 and an accuracy of 0.74.</p><p><strong>Conclusion: </strong>Predicting the PhA of inpatients from facial images is feasible and can be used for malnutrition assessment and prognostic prediction.</p>","PeriodicalId":19354,"journal":{"name":"Nutrition in Clinical Practice","volume":null,"pages":null},"PeriodicalIF":2.1000,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nutrition in Clinical Practice","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1002/ncp.11215","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"NUTRITION & DIETETICS","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Phase angle (PhA) correlates with body composition and could predict the nutrition status of patients and disease prognosis. We aimed to explore the feasibility of predicting PhA-diagnosed malnutrition using facial image information based on deep learning (DL).

Methods: From August 2021 to April 2022, inpatients were enrolled from surgery, gastroenterology, and oncology departments in a tertiary hospital. Subjective global assessment was used as the gold standard of malnutrition diagnosis. The highest Youden index value was selected as the PhA cutoff point. We developed a multimodal DL framework to automatically analyze the three-dimensional (3D) facial data and accurately determine patients' PhA categories. The framework was trained and validated using a cross-validation approach and tested on an independent dataset.

Results: Four hundred eighty-two patients were included in the final dataset, including 176 with malnourishment. In male patients, the PhA value with the highest Youden index was 5.55°, and the area under the receiver operating characteristic curve (AUC) = 0.68; in female patients, the PhA value with the highest Youden index was 4.88°, and AUC = 0.69. Inpatients with low PhA had higher incidence of infectious complications during the hospital stay (P = 0.003). The DL model trained with 4096 points extracted from 3D facial data had the best performance. The algorithm showed fair performance in predicting PhA, with an AUC of 0.77 and an accuracy of 0.74.

Conclusion: Predicting the PhA of inpatients from facial images is feasible and can be used for malnutrition assessment and prognostic prediction.

利用三维面部信息预测营养不良及其并发症。
背景:相位角(PhA)与身体成分相关,可以预测患者的营养状况和疾病预后。我们旨在探索基于深度学习(DL)的面部图像信息预测PhA诊断为营养不良的可行性:方法:2021 年 8 月至 2022 年 4 月,我们从一家三甲医院的外科、消化科和肿瘤科招募住院患者。主观综合评估作为营养不良诊断的金标准。尤登指数的最高值被选为 PhA 临界点。我们开发了一个多模态 DL 框架,用于自动分析三维(3D)面部数据并准确确定患者的 PhA 类别。我们采用交叉验证的方法对该框架进行了训练和验证,并在一个独立的数据集上进行了测试:最终数据集包括 482 名患者,其中包括 176 名营养不良患者。在男性患者中,尤登指数最高的 PhA 值为 5.55°,接收器工作特征曲线下面积(AUC)= 0.68;在女性患者中,尤登指数最高的 PhA 值为 4.88°,接收器工作特征曲线下面积(AUC)= 0.69。PhA 值低的住院患者在住院期间感染并发症的发生率更高(P = 0.003)。使用从三维面部数据中提取的 4096 个点训练的 DL 模型性能最佳。该算法在预测PhA方面表现尚可,AUC为0.77,准确率为0.74:通过面部图像预测住院患者的PhA是可行的,可用于营养不良评估和预后预测。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
6.00
自引率
9.70%
发文量
128
审稿时长
3 months
期刊介绍: NCP is a peer-reviewed, interdisciplinary publication that publishes articles about the scientific basis and clinical application of nutrition and nutrition support. NCP contains comprehensive reviews, clinical research, case observations, and other types of papers written by experts in the field of nutrition and health care practitioners involved in the delivery of specialized nutrition support. This journal is a member of the Committee on Publication Ethics (COPE).
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信