Mingyuan Qin , Lei Feng , Jing Lu , Ziyan Sun , Zhengyu Yu , Lianyi Han
{"title":"ZeroTuneBio NER:一个使用大型语言模型和快速工程的三阶段框架,用于零射击和零调谐生物医学实体提取","authors":"Mingyuan Qin , Lei Feng , Jing Lu , Ziyan Sun , Zhengyu Yu , Lianyi Han","doi":"10.1016/j.cmpb.2025.109070","DOIUrl":null,"url":null,"abstract":"<div><h3>Objective</h3><div>This study aims to (1) enhance the performance of large language models (LLMs) in biomedical entity extraction, (2) investigate zero-shot named entity recognition (NER) capabilities without fine-tuning, and (3) compare the proposed framework with existing models and human annotation methods. Additionally, we analyze discrepancies between human and LLM-generated annotations to refine manual labeling processes for specialized datasets.</div></div><div><h3>Materials and Methods</h3><div>We propose <strong>ZeroTuneBio NER</strong>, a three-stage NER framework integrating chain-of-thought reasoning and prompt engineering. Evaluated on three public datasets (disease, chemistry, and gene), the method requires no task-specific examples or LLM fine-tuning, addressing challenges in complex concept interpretation.</div></div><div><h3>Results</h3><div>ZeroTuneBio NER excels in tasks without strict matching, achieving an average F1-score improvement of <strong>0.28</strong> over direct LLM queries and a partial-matching F1-score of <strong>∼88</strong> <strong>%</strong>. It rivals the performance of a fine-tuned LLaMA model trained on <strong>11,240 examples</strong> and surpasses BioBERT trained on <strong>22,480 examples</strong> when strict-matching errors are excluded. Notably, LLMs significantly optimize manual annotation, accelerating speed and reducing costs.</div></div><div><h3>Conclusion</h3><div>ZeroTuneBio NER demonstrates that LLMs can perform high-quality NER without fine-tuning, reducing reliance on manual annotation. The framework broadens LLM applications in biomedical NER, while our analysis highlights its scalability and future research directions.</div></div>","PeriodicalId":10624,"journal":{"name":"Computer methods and programs in biomedicine","volume":"272 ","pages":"Article 109070"},"PeriodicalIF":4.8000,"publicationDate":"2025-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ZeroTuneBio NER: A three-stage framework for zero-shot and zero-tuning biomedical entity extraction using large language models and prompt engineering\",\"authors\":\"Mingyuan Qin , Lei Feng , Jing Lu , Ziyan Sun , Zhengyu Yu , Lianyi Han\",\"doi\":\"10.1016/j.cmpb.2025.109070\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Objective</h3><div>This study aims to (1) enhance the performance of large language models (LLMs) in biomedical entity extraction, (2) investigate zero-shot named entity recognition (NER) capabilities without fine-tuning, and (3) compare the proposed framework with existing models and human annotation methods. Additionally, we analyze discrepancies between human and LLM-generated annotations to refine manual labeling processes for specialized datasets.</div></div><div><h3>Materials and Methods</h3><div>We propose <strong>ZeroTuneBio NER</strong>, a three-stage NER framework integrating chain-of-thought reasoning and prompt engineering. Evaluated on three public datasets (disease, chemistry, and gene), the method requires no task-specific examples or LLM fine-tuning, addressing challenges in complex concept interpretation.</div></div><div><h3>Results</h3><div>ZeroTuneBio NER excels in tasks without strict matching, achieving an average F1-score improvement of <strong>0.28</strong> over direct LLM queries and a partial-matching F1-score of <strong>∼88</strong> <strong>%</strong>. It rivals the performance of a fine-tuned LLaMA model trained on <strong>11,240 examples</strong> and surpasses BioBERT trained on <strong>22,480 examples</strong> when strict-matching errors are excluded. Notably, LLMs significantly optimize manual annotation, accelerating speed and reducing costs.</div></div><div><h3>Conclusion</h3><div>ZeroTuneBio NER demonstrates that LLMs can perform high-quality NER without fine-tuning, reducing reliance on manual annotation. The framework broadens LLM applications in biomedical NER, while our analysis highlights its scalability and future research directions.</div></div>\",\"PeriodicalId\":10624,\"journal\":{\"name\":\"Computer methods and programs in biomedicine\",\"volume\":\"272 \",\"pages\":\"Article 109070\"},\"PeriodicalIF\":4.8000,\"publicationDate\":\"2025-09-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer methods and programs in biomedicine\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0169260725004870\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer methods and programs in biomedicine","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0169260725004870","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
ZeroTuneBio NER: A three-stage framework for zero-shot and zero-tuning biomedical entity extraction using large language models and prompt engineering
Objective
This study aims to (1) enhance the performance of large language models (LLMs) in biomedical entity extraction, (2) investigate zero-shot named entity recognition (NER) capabilities without fine-tuning, and (3) compare the proposed framework with existing models and human annotation methods. Additionally, we analyze discrepancies between human and LLM-generated annotations to refine manual labeling processes for specialized datasets.
Materials and Methods
We propose ZeroTuneBio NER, a three-stage NER framework integrating chain-of-thought reasoning and prompt engineering. Evaluated on three public datasets (disease, chemistry, and gene), the method requires no task-specific examples or LLM fine-tuning, addressing challenges in complex concept interpretation.
Results
ZeroTuneBio NER excels in tasks without strict matching, achieving an average F1-score improvement of 0.28 over direct LLM queries and a partial-matching F1-score of ∼88%. It rivals the performance of a fine-tuned LLaMA model trained on 11,240 examples and surpasses BioBERT trained on 22,480 examples when strict-matching errors are excluded. Notably, LLMs significantly optimize manual annotation, accelerating speed and reducing costs.
Conclusion
ZeroTuneBio NER demonstrates that LLMs can perform high-quality NER without fine-tuning, reducing reliance on manual annotation. The framework broadens LLM applications in biomedical NER, while our analysis highlights its scalability and future research directions.
期刊介绍:
To encourage the development of formal computing methods, and their application in biomedical research and medical practice, by illustration of fundamental principles in biomedical informatics research; to stimulate basic research into application software design; to report the state of research of biomedical information processing projects; to report new computer methodologies applied in biomedical areas; the eventual distribution of demonstrable software to avoid duplication of effort; to provide a forum for discussion and improvement of existing software; to optimize contact between national organizations and regional user groups by promoting an international exchange of information on formal methods, standards and software in biomedicine.
Computer Methods and Programs in Biomedicine covers computing methodology and software systems derived from computing science for implementation in all aspects of biomedical research and medical practice. It is designed to serve: biochemists; biologists; geneticists; immunologists; neuroscientists; pharmacologists; toxicologists; clinicians; epidemiologists; psychiatrists; psychologists; cardiologists; chemists; (radio)physicists; computer scientists; programmers and systems analysts; biomedical, clinical, electrical and other engineers; teachers of medical informatics and users of educational software.