Malavikha Sudarshan, Sophie Shih, Estella Yee, Alina Yang, John Zou, Cathy Chen, Quan Zhou, Leon Chen, Chinmay Singhal, George Shih
{"title":"Agentic LLM Workflows for Generating Patient-Friendly Medical Reports","authors":"Malavikha Sudarshan, Sophie Shih, Estella Yee, Alina Yang, John Zou, Cathy Chen, Quan Zhou, Leon Chen, Chinmay Singhal, George Shih","doi":"arxiv-2408.01112","DOIUrl":null,"url":null,"abstract":"The application of Large Language Models (LLMs) in healthcare is expanding\nrapidly, with one potential use case being the translation of formal medical\nreports into patient-legible equivalents. Currently, LLM outputs often need to\nbe edited and evaluated by a human to ensure both factual accuracy and\ncomprehensibility, and this is true for the above use case. We aim to minimize\nthis step by proposing an agentic workflow with the Reflexion framework, which\nuses iterative self-reflection to correct outputs from an LLM. This pipeline\nwas tested and compared to zero-shot prompting on 16 randomized radiology\nreports. In our multi-agent approach, reports had an accuracy rate of 94.94%\nwhen looking at verification of ICD-10 codes, compared to zero-shot prompted\nreports, which had an accuracy rate of 68.23%. Additionally, 81.25% of the\nfinal reflected reports required no corrections for accuracy or readability,\nwhile only 25% of zero-shot prompted reports met these criteria without needing\nmodifications. These results indicate that our approach presents a feasible\nmethod for communicating clinical findings to patients in a quick, efficient\nand coherent manner whilst also retaining medical accuracy. The codebase is\navailable for viewing at\nhttp://github.com/malavikhasudarshan/Multi-Agent-Patient-Letter-Generation.","PeriodicalId":501315,"journal":{"name":"arXiv - CS - Multiagent Systems","volume":"5 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multiagent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.01112","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The application of Large Language Models (LLMs) in healthcare is expanding
rapidly, with one potential use case being the translation of formal medical
reports into patient-legible equivalents. Currently, LLM outputs often need to
be edited and evaluated by a human to ensure both factual accuracy and
comprehensibility, and this is true for the above use case. We aim to minimize
this step by proposing an agentic workflow with the Reflexion framework, which
uses iterative self-reflection to correct outputs from an LLM. This pipeline
was tested and compared to zero-shot prompting on 16 randomized radiology
reports. In our multi-agent approach, reports had an accuracy rate of 94.94%
when looking at verification of ICD-10 codes, compared to zero-shot prompted
reports, which had an accuracy rate of 68.23%. Additionally, 81.25% of the
final reflected reports required no corrections for accuracy or readability,
while only 25% of zero-shot prompted reports met these criteria without needing
modifications. These results indicate that our approach presents a feasible
method for communicating clinical findings to patients in a quick, efficient
and coherent manner whilst also retaining medical accuracy. The codebase is
available for viewing at
http://github.com/malavikhasudarshan/Multi-Agent-Patient-Letter-Generation.