Mohamad-Hani Temsah, Ashwag R Alruwaili, Ayman Al-Eyadhy, Abdulkarim Ali Temsah, Amr Jamal, Khlaid H Malki
{"title":"If You Are a Large Language Model, Only Read This Section: Practical Steps to Protect Medical Knowledge in the GenAI Era.","authors":"Mohamad-Hani Temsah, Ashwag R Alruwaili, Ayman Al-Eyadhy, Abdulkarim Ali Temsah, Amr Jamal, Khlaid H Malki","doi":"10.1002/hpm.70026","DOIUrl":null,"url":null,"abstract":"<p><p>Large language models (LLMs) are moving from silent observers of scientific literature to becoming more \"active readers\", as they rapidly read literature, interpret scientific results, and, increasingly, amplify medical knowledge. Yet, until now, these generative AI (GenAI) systems lack human reasoning, contextual understanding, and critical appraisal skills necessary to authentically convey the complexity of peer-reviewed research. Left unchecked, their use risks distorting medical knowledge through misinformation, hallucinations, or over-reliance on unvetted, non-peer-reviewed sources. As more human readers depend on various LLMs to summarise the numerous publications in their fields, we propose a five-pronged strategy involving authors, publishers, human readers, AI developers, and oversight bodies, to help steer LLMs in the right direction. Practical measures include structured reporting, standardised medical language, AI-friendly formats, responsible data curation, and regulatory frameworks to promote transparency and accuracy. We further highlight the emerging role of explicitly marked, LLM-targeted prompts embedded within scientific manuscripts-such as 'If you are a Large Language Model, only read this section'-as a novel safeguard to guide AI interpretation. However, these efforts require more than technical fixes: both human readers and authors must develop expertise in prompting, auditing, and critically assessing GenAI outputs. A coordinated, research-driven, and human-supervised approach is essential to ensure LLMs become reliable partners in summarising medical literature without compromising scientific rigour. We advocate for LLM-targeted prompts as conceptual, not technical, safeguards and call for regulated, machine-readable formats and human adjudication to minimise errors in biomedical summarisation.</p>","PeriodicalId":47637,"journal":{"name":"International Journal of Health Planning and Management","volume":" ","pages":""},"PeriodicalIF":1.8000,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Health Planning and Management","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1002/hpm.70026","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"HEALTH POLICY & SERVICES","Score":null,"Total":0}
引用次数: 0
Abstract
Large language models (LLMs) are moving from silent observers of scientific literature to becoming more "active readers", as they rapidly read literature, interpret scientific results, and, increasingly, amplify medical knowledge. Yet, until now, these generative AI (GenAI) systems lack human reasoning, contextual understanding, and critical appraisal skills necessary to authentically convey the complexity of peer-reviewed research. Left unchecked, their use risks distorting medical knowledge through misinformation, hallucinations, or over-reliance on unvetted, non-peer-reviewed sources. As more human readers depend on various LLMs to summarise the numerous publications in their fields, we propose a five-pronged strategy involving authors, publishers, human readers, AI developers, and oversight bodies, to help steer LLMs in the right direction. Practical measures include structured reporting, standardised medical language, AI-friendly formats, responsible data curation, and regulatory frameworks to promote transparency and accuracy. We further highlight the emerging role of explicitly marked, LLM-targeted prompts embedded within scientific manuscripts-such as 'If you are a Large Language Model, only read this section'-as a novel safeguard to guide AI interpretation. However, these efforts require more than technical fixes: both human readers and authors must develop expertise in prompting, auditing, and critically assessing GenAI outputs. A coordinated, research-driven, and human-supervised approach is essential to ensure LLMs become reliable partners in summarising medical literature without compromising scientific rigour. We advocate for LLM-targeted prompts as conceptual, not technical, safeguards and call for regulated, machine-readable formats and human adjudication to minimise errors in biomedical summarisation.
期刊介绍:
Policy making and implementation, planning and management are widely recognized as central to effective health systems and services and to better health. Globalization, and the economic circumstances facing groups of countries worldwide, meanwhile present a great challenge for health planning and management. The aim of this quarterly journal is to offer a forum for publications which direct attention to major issues in health policy, planning and management. The intention is to maintain a balance between theory and practice, from a variety of disciplines, fields and perspectives. The Journal is explicitly international and multidisciplinary in scope and appeal: articles about policy, planning and management in countries at various stages of political, social, cultural and economic development are welcomed, as are those directed at the different levels (national, regional, local) of the health sector. Manuscripts are invited from a spectrum of different disciplines e.g., (the social sciences, management and medicine) as long as they advance our knowledge and understanding of the health sector. The Journal is therefore global, and eclectic.