Emotional prompting amplifies disinformation generation in AI large language models.

IF 3 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Frontiers in Artificial Intelligence Pub Date : 2025-04-07 eCollection Date: 2025-01-01 DOI:10.3389/frai.2025.1543603
Rasita Vinay, Giovanni Spitale, Nikola Biller-Andorno, Federico Germani
{"title":"Emotional prompting amplifies disinformation generation in AI large language models.","authors":"Rasita Vinay, Giovanni Spitale, Nikola Biller-Andorno, Federico Germani","doi":"10.3389/frai.2025.1543603","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>The emergence of artificial intelligence (AI) large language models (LLMs), which can produce text that closely resembles human-written content, presents both opportunities and risks. While these developments offer significant opportunities for improving communication, such as in health-related crisis communication, they also pose substantial risks by facilitating the creation of convincing fake news and disinformation. The widespread dissemination of AI-generated disinformation adds complexity to the existing challenges of the ongoing infodemic, significantly affecting public health and the stability of democratic institutions.</p><p><strong>Rationale: </strong>Prompt engineering is a technique that involves the creation of specific queries given to LLMs. It has emerged as a strategy to guide LLMs in generating the desired outputs. Recent research shows that the output of LLMs depends on emotional framing within prompts, suggesting that incorporating emotional cues into prompts could influence their response behavior. In this study, we investigated how the politeness or impoliteness of prompts affects the frequency of disinformation generation by various LLMs.</p><p><strong>Results: </strong>We generated and evaluated a corpus of 19,800 social media posts on public health topics to assess the disinformation generation capabilities of OpenAI's LLMs, including davinci-002, davinci-003, gpt-3.5-turbo, and gpt-4. Our findings revealed that all LLMs efficiently generated disinformation (davinci-002, 67%; davinci-003, 86%; gpt-3.5-turbo, 77%; and gpt-4, 99%). Introducing polite language to prompt requests yielded significantly higher success rates for disinformation (davinci-002, 79%; davinci-003, 90%; gpt-3.5-turbo, 94%; and gpt-4, 100%). Impolite prompting resulted in a significant decrease in disinformation production across all models (davinci-002, 59%; davinci-003, 44%; and gpt-3.5-turbo, 28%) and a slight reduction for gpt-4 (94%).</p><p><strong>Conclusion: </strong>Our study reveals that all tested LLMs effectively generate disinformation. Notably, emotional prompting had a significant impact on disinformation production rates, with models showing higher success rates when prompted with polite language compared to neutral or impolite requests. Our investigation highlights that LLMs can be exploited to create disinformation and emphasizes the critical need for ethics-by-design approaches in developing AI technologies. We maintain that identifying ways to mitigate the exploitation of LLMs through emotional prompting is crucial to prevent their misuse for purposes detrimental to public health and society.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"8 ","pages":"1543603"},"PeriodicalIF":3.0000,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12009909/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/frai.2025.1543603","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction: The emergence of artificial intelligence (AI) large language models (LLMs), which can produce text that closely resembles human-written content, presents both opportunities and risks. While these developments offer significant opportunities for improving communication, such as in health-related crisis communication, they also pose substantial risks by facilitating the creation of convincing fake news and disinformation. The widespread dissemination of AI-generated disinformation adds complexity to the existing challenges of the ongoing infodemic, significantly affecting public health and the stability of democratic institutions.

Rationale: Prompt engineering is a technique that involves the creation of specific queries given to LLMs. It has emerged as a strategy to guide LLMs in generating the desired outputs. Recent research shows that the output of LLMs depends on emotional framing within prompts, suggesting that incorporating emotional cues into prompts could influence their response behavior. In this study, we investigated how the politeness or impoliteness of prompts affects the frequency of disinformation generation by various LLMs.

Results: We generated and evaluated a corpus of 19,800 social media posts on public health topics to assess the disinformation generation capabilities of OpenAI's LLMs, including davinci-002, davinci-003, gpt-3.5-turbo, and gpt-4. Our findings revealed that all LLMs efficiently generated disinformation (davinci-002, 67%; davinci-003, 86%; gpt-3.5-turbo, 77%; and gpt-4, 99%). Introducing polite language to prompt requests yielded significantly higher success rates for disinformation (davinci-002, 79%; davinci-003, 90%; gpt-3.5-turbo, 94%; and gpt-4, 100%). Impolite prompting resulted in a significant decrease in disinformation production across all models (davinci-002, 59%; davinci-003, 44%; and gpt-3.5-turbo, 28%) and a slight reduction for gpt-4 (94%).

Conclusion: Our study reveals that all tested LLMs effectively generate disinformation. Notably, emotional prompting had a significant impact on disinformation production rates, with models showing higher success rates when prompted with polite language compared to neutral or impolite requests. Our investigation highlights that LLMs can be exploited to create disinformation and emphasizes the critical need for ethics-by-design approaches in developing AI technologies. We maintain that identifying ways to mitigate the exploitation of LLMs through emotional prompting is crucial to prevent their misuse for purposes detrimental to public health and society.

在人工智能大型语言模型中,情绪提示放大了虚假信息的产生。
导读:人工智能(AI)大型语言模型(llm)的出现,可以产生与人类编写的内容非常相似的文本,这既带来了机遇,也带来了风险。虽然这些发展为改善沟通提供了重要机会,例如在与健康有关的危机沟通方面,但它们也带来了重大风险,因为它们促进了令人信服的假新闻和虚假信息的产生。人工智能产生的虚假信息的广泛传播使当前信息流行的现有挑战更加复杂,严重影响公共卫生和民主体制的稳定。基本原理:提示工程是一种技术,它涉及到创建给定给llm的特定查询。它已成为指导法学硕士产生所需产出的战略。最近的研究表明,llm的输出依赖于提示中的情感框架,这表明将情感线索纳入提示可能会影响他们的反应行为。在本研究中,我们调查了提示语的礼貌或不礼貌如何影响各种法学硕士产生虚假信息的频率。结果:我们生成并评估了关于公共卫生主题的19,800个社交媒体帖子的语料库,以评估OpenAI llm的虚假信息生成能力,包括davinci-002, davinci-003, gpt-3.5-turbo和gpt-4。我们的研究结果显示,所有法学硕士都有效地制造了虚假信息(davinci-002, 67%;达芬奇- 003、86%;gpt - 3.5 -涡轮,77%;gpt-4占99%)。在提示请求中引入礼貌用语可以显著提高虚假信息的成功率(davinci-002, 79%;达芬奇- 003、90%;gpt - 3.5 -涡轮,94%;gpt-4, 100%)。不礼貌的提示导致所有模型中虚假信息的产生显著减少(davinci-002, 59%;达芬奇- 003、44%;gpt-3.5涡轮,28%),gpt-4略有下降(94%)。结论:我们的研究表明,所有被测试的法学硕士都有效地制造了虚假信息。值得注意的是,情绪提示对虚假信息的产生率有显著影响,与中性或不礼貌的请求相比,用礼貌语言提示的模型显示出更高的成功率。我们的调查强调了法学硕士可以被利用来制造虚假信息,并强调了在开发人工智能技术时对设计伦理方法的迫切需要。我们认为,确定通过情感激励来减少法学硕士被利用的方法,对于防止法学硕士被滥用于有害公共健康和社会的目的至关重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
6.10
自引率
2.50%
发文量
272
审稿时长
13 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信