{"title":"提示式宣传:追踪大型语言模型中隐藏的语言策略","authors":"Arash Barfar, Lee Sommerfeldt","doi":"10.1016/j.ipm.2025.104403","DOIUrl":null,"url":null,"abstract":"<div><div>As large language models become increasingly integrated into news production, concerns have grown over their potential to generate polarizing propaganda. This study introduces a scalable and flexible framework for systematically tracing the rhetorical strategies LLMs use to produce propaganda-style content. We apply the framework across three versions of GPT (GPT-3.5-Turbo, GPT-4o, and GPT-4.1), generating over 340,000 articles on selected politically divisive topics in the American news landscape. Supported by highly consistent distinctions (AUROC above 98 %), our findings reveal that the persuasive strategies adopted by GPT are both coherent and evolving across model versions. All three models rely heavily on cognitive language to simulate deliberation and interpretive reasoning, combined with consistent use of moral framing. Each version layers this rhetorical core with distinct stylistic choices: GPT-3.5-Turbo emphasizes collective identity and narrative looseness; GPT-4o adopts reflective detachment through its use of impersonal pronouns and tentative language; and GPT-4.1 deploys lexical sophistication and definitive assertions to project authority. These differences reflect a rhetorical evolution driven by architectural refinements, training updates, and changes in safety guard behavior. A comparison with human-authored propaganda further shows that GPT is not simply reproducing prompt-induced rhetorical biases but appears to exhibit distinct generative tendencies beyond those present in the human-authored baselines. The framework developed here offers a practical reverse-engineering tool for researchers, policymakers, and developers to explain and audit the persuasive capabilities of LLMs. It contributes to broader efforts in AI transparency, content moderation, and the promotion of epistemic resilience in digital communication.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 2","pages":"Article 104403"},"PeriodicalIF":6.9000,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Propaganda by prompt: Tracing hidden linguistic strategies in large language models\",\"authors\":\"Arash Barfar, Lee Sommerfeldt\",\"doi\":\"10.1016/j.ipm.2025.104403\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>As large language models become increasingly integrated into news production, concerns have grown over their potential to generate polarizing propaganda. This study introduces a scalable and flexible framework for systematically tracing the rhetorical strategies LLMs use to produce propaganda-style content. We apply the framework across three versions of GPT (GPT-3.5-Turbo, GPT-4o, and GPT-4.1), generating over 340,000 articles on selected politically divisive topics in the American news landscape. Supported by highly consistent distinctions (AUROC above 98 %), our findings reveal that the persuasive strategies adopted by GPT are both coherent and evolving across model versions. All three models rely heavily on cognitive language to simulate deliberation and interpretive reasoning, combined with consistent use of moral framing. Each version layers this rhetorical core with distinct stylistic choices: GPT-3.5-Turbo emphasizes collective identity and narrative looseness; GPT-4o adopts reflective detachment through its use of impersonal pronouns and tentative language; and GPT-4.1 deploys lexical sophistication and definitive assertions to project authority. These differences reflect a rhetorical evolution driven by architectural refinements, training updates, and changes in safety guard behavior. A comparison with human-authored propaganda further shows that GPT is not simply reproducing prompt-induced rhetorical biases but appears to exhibit distinct generative tendencies beyond those present in the human-authored baselines. The framework developed here offers a practical reverse-engineering tool for researchers, policymakers, and developers to explain and audit the persuasive capabilities of LLMs. It contributes to broader efforts in AI transparency, content moderation, and the promotion of epistemic resilience in digital communication.</div></div>\",\"PeriodicalId\":50365,\"journal\":{\"name\":\"Information Processing & Management\",\"volume\":\"63 2\",\"pages\":\"Article 104403\"},\"PeriodicalIF\":6.9000,\"publicationDate\":\"2025-09-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Processing & Management\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306457325003449\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325003449","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Propaganda by prompt: Tracing hidden linguistic strategies in large language models
As large language models become increasingly integrated into news production, concerns have grown over their potential to generate polarizing propaganda. This study introduces a scalable and flexible framework for systematically tracing the rhetorical strategies LLMs use to produce propaganda-style content. We apply the framework across three versions of GPT (GPT-3.5-Turbo, GPT-4o, and GPT-4.1), generating over 340,000 articles on selected politically divisive topics in the American news landscape. Supported by highly consistent distinctions (AUROC above 98 %), our findings reveal that the persuasive strategies adopted by GPT are both coherent and evolving across model versions. All three models rely heavily on cognitive language to simulate deliberation and interpretive reasoning, combined with consistent use of moral framing. Each version layers this rhetorical core with distinct stylistic choices: GPT-3.5-Turbo emphasizes collective identity and narrative looseness; GPT-4o adopts reflective detachment through its use of impersonal pronouns and tentative language; and GPT-4.1 deploys lexical sophistication and definitive assertions to project authority. These differences reflect a rhetorical evolution driven by architectural refinements, training updates, and changes in safety guard behavior. A comparison with human-authored propaganda further shows that GPT is not simply reproducing prompt-induced rhetorical biases but appears to exhibit distinct generative tendencies beyond those present in the human-authored baselines. The framework developed here offers a practical reverse-engineering tool for researchers, policymakers, and developers to explain and audit the persuasive capabilities of LLMs. It contributes to broader efforts in AI transparency, content moderation, and the promotion of epistemic resilience in digital communication.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.