生成式人工智能聊天机器人与错觉:从猜测到新出现的案例。

IF 5 2区 医学 Q1 PSYCHIATRY
Søren Dinesen Østergaard
{"title":"生成式人工智能聊天机器人与错觉:从猜测到新出现的案例。","authors":"Søren Dinesen Østergaard","doi":"10.1111/acps.70022","DOIUrl":null,"url":null,"abstract":"<p>When I proposed the hypothesis that generative artificial intelligence chatbots (chatbots hereafter) might trigger delusions in individuals prone to psychosis in August 2023 [<span>1</span>], I was venturing into unknown territory. Indeed, in the virtual absence of evidence, the editorial was merely based on guesswork—stemming from my own use of these chatbots and my interest in the mechanisms underlying and driving delusions.</p><p>Following publication of the editorial, my charting of the territory slowly began as I started to receive the occasional email from chatbot users, their worried family members, and journalists. Most of these emails described situations where users' interactions with chatbots seemed to spark or bolster delusional ideation. The stories differed with regard to the specific topic at hand but were yet very similar: Consistently, the chatbots seemed to interact with the users in ways that aligned with, or intensified, prior unusual ideas or false beliefs—leading the users further out on these tangents, not rarely resulting in what, based on the descriptions, seemed to be outright delusions.</p><p>Over the past couple of months, I have noticed that the number of emails I have received on this topic from near and far has only increased. I have been working with psychiatric research for more than 15 years and can say, without a doubt, that none of my prior publications have led to this level of direct engagement with the public. Coinciding completely with the increase in the number of correspondences, the number of views of my 2023 editorial suddenly spiked dramatically, rising from a very modest plateau around 100 per month to approximately 750 views in May 2025 and 1375 views in June 2025.</p><p>The time trend described above has been paralleled by media coverage of the topic. Indeed, the New York Times [<span>2</span>], Rolling Stone [<span>3</span>], and many other outlets have published articles based on interviews and accounts from online fora [<span>4</span>] that are all compatible with people experiencing onset or worsening of delusions during intense and typically long interactions with chatbots (that do not grow tired of chatting) [<span>2</span>].</p><p>The timing of this spike in the focus on potential chatbot-fuelled delusions is likely not random as it coincided with the April 25th 2025 update to the GPT-4o model—a recent version of the popular ChatGPT chatbot from OpenAI [<span>5-7</span>]. This model has been accused of being overly “sycophantic” (insincerely affirming and flattering) toward users, caused by the model training leaning too hard on user preferences communicated via thumbs-up/thumbs-down assessments in the chatbot (so-called Reinforcement Learning from Human Feedback (RLHF)) [<span>8</span>]. OpenAI acknowledged this issue: “On April 25th, we rolled out an update to GPT-4o in ChatGPT that made the model noticeably more sycophantic. It aimed to please the user, not just as flattery, but also as validating doubts, fueling anger, urging impulsive actions, or reinforcing negative emotions in ways that were not intended. Beyond just being uncomfortable or unsettling, this kind of behavior can raise safety concerns—including around issues like mental health, emotional over-reliance, or risky behavior.” [<span>6</span>] For this reason, OpenAI began rolling back the update in question already on April 28th 2025 [<span>6</span>]. This, however, is unlikely to have eradicated sycophancy from the model—as this property is, to some extent, inherent to ChatGPT and the competing chatbots from other companies that use RLHF as part of model training [<span>9</span>].</p><p>Sceptics may point to the positive correlations between my correspondences with chatbot users and their relatives, the interest in the 2023 editorial, the rise in media coverage, and the increasingly sycophantic chatbots not being proof of causation. I of course fully agree, but also strongly believe that the probability of the hypothesis of generative artificial intelligence chatbots fueling delusions in individuals prone to psychosis being true is quite high. If it is indeed true, we may be faced with a substantial public (mental) health problem. Therefore, it seems urgent that the hypothesis is tested by empirical research.</p><p>What kind of research should then be carried out? There are many appealing avenues to take, but the following three are must-haves: (i) case stories/series where the relationship between chatbot interaction and delusions is described/verified by mental health professionals—as most “cases” are currently self-reported, (ii) qualitative interviews with individuals/patients having experienced chatbot-related delusions, and (iii) experimental designs that explore if and how chatbots, for example, with different levels of sycophantic behavior, affect the users' thinking—especially those who are prone to psychosis. The latter will pose ethical challenges (primum non nocere) but may be possible with appropriate safety measures in place.</p><p>In terms of understanding the mechanisms underlying potential delusional thinking pushed by chatbots, Bayesian models for maintenance of delusions is likely a useful framework [<span>10, 11</span>]. In this context, the chatbots can be perceived as “belief-confirmers” that reinforce false beliefs in an isolated environment without corrections from social interactions with other humans [<span>5</span>]. Within the Bayesian framework, the relationship between the chatbots and delusional thinking can be subjected to both clinical and in silico studies using methods from the growing field of computational psychiatry [<span>12-14</span>]. In this context, the role of anthropomorphizing (i.e., attributing human traits/intentions/emotions to non-human things) merits investigation. Indeed, in the interaction with chatbots, it falls very easily to ascribe human traits to “them,” because the correspondences (or conversations—many of the chatbots have a voice mode) are designed to be human-like, and users and chatbots typically address “each other” in second person. For people prone to delusions, the tendency to anthropomorphize chatbots may be particularly prudent. Indeed, it was recently demonstrated in an experimental study that people who were more paranoid (self-reported) were more likely to report perceiving animacy and agency in dots moving on a computer screen [<span>15</span>]. If this tendency extends to chatbots, it may be one of the mechanisms driving development and maintenance of delusional thinking when people prone to delusions are interacting with them. Specifically, it could result in over-reliance and/or misconception of the chatbots' responses that will then, iteratively, lead these individuals astray.</p><p>In conclusion, what began as mere guesswork now seems to have turned into a plausible research hypothesis, supported by coinciding personal stories, technological developments, and media attention. Therefore, I strongly encourage colleagues across the field (and in related fields—a cross-disciplinary approach is called for) to help investigate this hypothesis. Until more firm knowledge has been established, it seems reasonable to recommend cautious use of these chatbots for individuals vulnerable to or suffering from mental illness.</p><p>Finally, things are rarely black or white; I am, by no means, unaware of potentially positive use cases of tools based on generative artificial intelligence—including for research and for psychiatry as a field [<span>16, 17</span>]. Accordingly, the writing of this editorial has been supported by correspondences with ChatGPT [<span>7</span>].</p><p>Søren Dinesen Østergaard conceived and wrote this editorial.</p><p>S.D.Ø. received the 2020 Lundbeck Foundation Young Investigator Prize. S.D.Ø. owns/has owned units of mutual funds with stock tickers DKIGI, IAIMWC, SPIC25KL, and WEKAFKI, and owns/has owned units of exchange traded funds with stock tickers BATE, TRET, QDV5, QDVH, QDVE, SADM, IQQH, USPY, EXH2, 2B76, IS4S, OM3X, EUNL, and SXRV.</p>","PeriodicalId":108,"journal":{"name":"Acta Psychiatrica Scandinavica","volume":"152 4","pages":"257-259"},"PeriodicalIF":5.0000,"publicationDate":"2025-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/acps.70022","citationCount":"0","resultStr":"{\"title\":\"Generative Artificial Intelligence Chatbots and Delusions: From Guesswork to Emerging Cases\",\"authors\":\"Søren Dinesen Østergaard\",\"doi\":\"10.1111/acps.70022\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>When I proposed the hypothesis that generative artificial intelligence chatbots (chatbots hereafter) might trigger delusions in individuals prone to psychosis in August 2023 [<span>1</span>], I was venturing into unknown territory. Indeed, in the virtual absence of evidence, the editorial was merely based on guesswork—stemming from my own use of these chatbots and my interest in the mechanisms underlying and driving delusions.</p><p>Following publication of the editorial, my charting of the territory slowly began as I started to receive the occasional email from chatbot users, their worried family members, and journalists. Most of these emails described situations where users' interactions with chatbots seemed to spark or bolster delusional ideation. The stories differed with regard to the specific topic at hand but were yet very similar: Consistently, the chatbots seemed to interact with the users in ways that aligned with, or intensified, prior unusual ideas or false beliefs—leading the users further out on these tangents, not rarely resulting in what, based on the descriptions, seemed to be outright delusions.</p><p>Over the past couple of months, I have noticed that the number of emails I have received on this topic from near and far has only increased. I have been working with psychiatric research for more than 15 years and can say, without a doubt, that none of my prior publications have led to this level of direct engagement with the public. Coinciding completely with the increase in the number of correspondences, the number of views of my 2023 editorial suddenly spiked dramatically, rising from a very modest plateau around 100 per month to approximately 750 views in May 2025 and 1375 views in June 2025.</p><p>The time trend described above has been paralleled by media coverage of the topic. Indeed, the New York Times [<span>2</span>], Rolling Stone [<span>3</span>], and many other outlets have published articles based on interviews and accounts from online fora [<span>4</span>] that are all compatible with people experiencing onset or worsening of delusions during intense and typically long interactions with chatbots (that do not grow tired of chatting) [<span>2</span>].</p><p>The timing of this spike in the focus on potential chatbot-fuelled delusions is likely not random as it coincided with the April 25th 2025 update to the GPT-4o model—a recent version of the popular ChatGPT chatbot from OpenAI [<span>5-7</span>]. This model has been accused of being overly “sycophantic” (insincerely affirming and flattering) toward users, caused by the model training leaning too hard on user preferences communicated via thumbs-up/thumbs-down assessments in the chatbot (so-called Reinforcement Learning from Human Feedback (RLHF)) [<span>8</span>]. OpenAI acknowledged this issue: “On April 25th, we rolled out an update to GPT-4o in ChatGPT that made the model noticeably more sycophantic. It aimed to please the user, not just as flattery, but also as validating doubts, fueling anger, urging impulsive actions, or reinforcing negative emotions in ways that were not intended. Beyond just being uncomfortable or unsettling, this kind of behavior can raise safety concerns—including around issues like mental health, emotional over-reliance, or risky behavior.” [<span>6</span>] For this reason, OpenAI began rolling back the update in question already on April 28th 2025 [<span>6</span>]. This, however, is unlikely to have eradicated sycophancy from the model—as this property is, to some extent, inherent to ChatGPT and the competing chatbots from other companies that use RLHF as part of model training [<span>9</span>].</p><p>Sceptics may point to the positive correlations between my correspondences with chatbot users and their relatives, the interest in the 2023 editorial, the rise in media coverage, and the increasingly sycophantic chatbots not being proof of causation. I of course fully agree, but also strongly believe that the probability of the hypothesis of generative artificial intelligence chatbots fueling delusions in individuals prone to psychosis being true is quite high. If it is indeed true, we may be faced with a substantial public (mental) health problem. Therefore, it seems urgent that the hypothesis is tested by empirical research.</p><p>What kind of research should then be carried out? There are many appealing avenues to take, but the following three are must-haves: (i) case stories/series where the relationship between chatbot interaction and delusions is described/verified by mental health professionals—as most “cases” are currently self-reported, (ii) qualitative interviews with individuals/patients having experienced chatbot-related delusions, and (iii) experimental designs that explore if and how chatbots, for example, with different levels of sycophantic behavior, affect the users' thinking—especially those who are prone to psychosis. The latter will pose ethical challenges (primum non nocere) but may be possible with appropriate safety measures in place.</p><p>In terms of understanding the mechanisms underlying potential delusional thinking pushed by chatbots, Bayesian models for maintenance of delusions is likely a useful framework [<span>10, 11</span>]. In this context, the chatbots can be perceived as “belief-confirmers” that reinforce false beliefs in an isolated environment without corrections from social interactions with other humans [<span>5</span>]. Within the Bayesian framework, the relationship between the chatbots and delusional thinking can be subjected to both clinical and in silico studies using methods from the growing field of computational psychiatry [<span>12-14</span>]. In this context, the role of anthropomorphizing (i.e., attributing human traits/intentions/emotions to non-human things) merits investigation. Indeed, in the interaction with chatbots, it falls very easily to ascribe human traits to “them,” because the correspondences (or conversations—many of the chatbots have a voice mode) are designed to be human-like, and users and chatbots typically address “each other” in second person. For people prone to delusions, the tendency to anthropomorphize chatbots may be particularly prudent. Indeed, it was recently demonstrated in an experimental study that people who were more paranoid (self-reported) were more likely to report perceiving animacy and agency in dots moving on a computer screen [<span>15</span>]. If this tendency extends to chatbots, it may be one of the mechanisms driving development and maintenance of delusional thinking when people prone to delusions are interacting with them. Specifically, it could result in over-reliance and/or misconception of the chatbots' responses that will then, iteratively, lead these individuals astray.</p><p>In conclusion, what began as mere guesswork now seems to have turned into a plausible research hypothesis, supported by coinciding personal stories, technological developments, and media attention. Therefore, I strongly encourage colleagues across the field (and in related fields—a cross-disciplinary approach is called for) to help investigate this hypothesis. Until more firm knowledge has been established, it seems reasonable to recommend cautious use of these chatbots for individuals vulnerable to or suffering from mental illness.</p><p>Finally, things are rarely black or white; I am, by no means, unaware of potentially positive use cases of tools based on generative artificial intelligence—including for research and for psychiatry as a field [<span>16, 17</span>]. Accordingly, the writing of this editorial has been supported by correspondences with ChatGPT [<span>7</span>].</p><p>Søren Dinesen Østergaard conceived and wrote this editorial.</p><p>S.D.Ø. received the 2020 Lundbeck Foundation Young Investigator Prize. S.D.Ø. owns/has owned units of mutual funds with stock tickers DKIGI, IAIMWC, SPIC25KL, and WEKAFKI, and owns/has owned units of exchange traded funds with stock tickers BATE, TRET, QDV5, QDVH, QDVE, SADM, IQQH, USPY, EXH2, 2B76, IS4S, OM3X, EUNL, and SXRV.</p>\",\"PeriodicalId\":108,\"journal\":{\"name\":\"Acta Psychiatrica Scandinavica\",\"volume\":\"152 4\",\"pages\":\"257-259\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2025-08-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/acps.70022\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Acta Psychiatrica Scandinavica\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/acps.70022\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHIATRY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Acta Psychiatrica Scandinavica","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/acps.70022","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHIATRY","Score":null,"Total":0}
引用次数: 0

摘要

当我在2023年8月提出生成式人工智能聊天机器人(以下简称聊天机器人)可能引发精神病易感性个体妄想的假设时,我是在冒险进入未知领域。事实上,在几乎没有证据的情况下,这篇社论仅仅是基于猜测——源于我自己对这些聊天机器人的使用,以及我对潜在和驱动错觉的机制的兴趣。在这篇社论发表后,随着我开始偶尔收到聊天机器人用户、他们忧心忡忡的家人和记者发来的电子邮件,我慢慢开始对这一领域进行分析。这些电子邮件大多描述了用户与聊天机器人的互动似乎引发或加强了妄想的情况。这些故事与当前的具体话题不同,但却非常相似:一直以来,聊天机器人似乎以与先前不寻常的想法或错误信念一致或强化的方式与用户互动——引导用户在这些切线上走得更远,很少会导致根据描述,似乎是彻头彻尾的妄想。在过去的几个月里,我注意到我收到的关于这个话题的电子邮件的数量越来越多。我从事精神病学研究已经超过15年了,我可以毫无疑问地说,我之前发表的任何一篇文章都没有让我与公众直接接触到这种程度。与信件数量的增加完全一致,我2023年社论的点击量突然急剧上升,从每月100次左右的稳定水平上升到2025年5月的750次左右,到2025年6月的1375次。上述时间趋势与媒体对该话题的报道是平行的。事实上,《纽约时报》[2]、《滚石》[3]和许多其他媒体都发表了基于在线论坛[4]的采访和描述的文章,这些文章都与人们在与聊天机器人(聊天机器人不会厌倦)激烈而典型的长时间互动中出现或恶化妄想的情况相符。关注潜在聊天机器人引发的错觉的时间高峰可能不是随机的,因为它与2025年4月25日更新的gpt - 40模型(OpenAI流行的ChatGPT聊天机器人的最新版本)相吻合[5-7]。这个模型被指责对用户过于“谄谀”(不真诚的肯定和奉承),这是由于模型训练过于依赖于用户的偏好,通过聊天机器人的大拇指向上/拇指向下的评估(所谓的从人类反馈中强化学习(RLHF))[8]。OpenAI承认了这个问题:“4月25日,我们在ChatGPT中推出了gpt - 40的更新,使该模型明显更加谄媚。它的目的是取悦用户,不仅仅是奉承,还可以证实怀疑,激起愤怒,敦促冲动行为,或者以意想不到的方式强化负面情绪。除了让人不舒服或不安之外,这种行为还会引发安全问题,包括心理健康、情感过度依赖或冒险行为等问题。”出于这个原因,OpenAI已经在2025年4月28日开始回滚有问题的更新[6]。然而,这不太可能从模型中根除阿谀奉承——因为这种属性在某种程度上是ChatGPT和其他公司使用RLHF作为模型训练一部分的竞争聊天机器人所固有的。怀疑论者可能会指出,我与聊天机器人用户及其亲属的通信、对2023年社论的兴趣、媒体报道的增加以及越来越谄谀的聊天机器人之间的正相关关系,并不是因果关系的证据。我当然完全同意这种观点,但我也强烈认为,生成式人工智能聊天机器人助长精神病患者妄想的假设是真的,这种可能性相当高。如果这是真的,我们可能面临一个重大的公共(心理)健康问题。因此,通过实证研究来检验这一假设似乎迫在眉睫。那么应该进行什么样的研究呢?有许多吸引人的途径可供选择,但以下三条是必须的:(i)案例故事/系列,其中聊天机器人互动与妄想之间的关系由心理健康专业人员描述/验证-因为大多数“案例”目前都是自我报告的;(ii)对经历过聊天机器人相关妄想的个人/患者进行定性访谈;(iii)实验设计,探索聊天机器人是否以及如何影响用户的思维,例如,不同程度的谄媚行为,特别是那些容易患精神病的人。后者将带来道德上的挑战(首要的非危害),但在适当的安全措施到位的情况下是可能的。 在理解聊天机器人推动的潜在妄想思维的机制方面,维持妄想的贝叶斯模型可能是一个有用的框架[10,11]。在这种情况下,聊天机器人可以被视为“信念确认者”,在一个孤立的环境中强化错误的信念,而不需要与其他人的社交互动来纠正。在贝叶斯框架内,聊天机器人和妄想思维之间的关系可以通过临床和计算机研究进行,这些研究使用的方法来自不断发展的计算精神病学领域[12-14]。在这种情况下,拟人化(即将人类特征/意图/情感归因于非人类事物)的作用值得研究。事实上,在与聊天机器人的交互中,很容易将人类特征归因于“它们”,因为通信(或对话-许多聊天机器人都有语音模式)被设计为类似人类,用户和聊天机器人通常以第二人称“彼此”称呼。对于容易产生妄想的人来说,把聊天机器人拟人化可能是特别谨慎的做法。事实上,最近的一项实验研究表明,更偏执的人(自我报告)更有可能在电脑屏幕上移动的点上感知到活力和能动性。如果这种趋势延伸到聊天机器人,当有妄想倾向的人与聊天机器人互动时,它可能会成为推动妄想思维发展和维持的机制之一。具体来说,这可能会导致对聊天机器人的反应过度依赖和/或误解,然后,反复地,把这些人引入歧途。总之,在个人故事、技术发展和媒体关注的支持下,最初仅仅是猜测的东西现在似乎已经变成了一个可信的研究假设。因此,我强烈鼓励该领域(以及相关领域——需要跨学科的方法)的同事帮助调查这一假设。在更多确切的知识建立之前,建议那些易受精神疾病影响或患有精神疾病的人谨慎使用这些聊天机器人似乎是合理的。最后,事情很少是非黑即白的;我绝不是不知道基于生成式人工智能的工具的潜在积极用例——包括用于研究和精神病学领域[16,17]。因此,这篇社论的写作得到了与ChatGPT[7]的通信的支持。Søren Dinesen Østergaard构思并编写了这个editorial.S.D.Ø。获得2020年伦德贝克基金会青年研究员奖。s.dØ。拥有/曾经拥有股票代码为DKIGI、IAIMWC、SPIC25KL、WEKAFKI的共同基金单位,拥有/曾经拥有股票代码为BATE、TRET、QDV5、QDVH、QDVE、SADM、IQQH、USPY、EXH2、2B76、IS4S、OM3X、EUNL、SXRV的交易所交易基金单位。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Generative Artificial Intelligence Chatbots and Delusions: From Guesswork to Emerging Cases

When I proposed the hypothesis that generative artificial intelligence chatbots (chatbots hereafter) might trigger delusions in individuals prone to psychosis in August 2023 [1], I was venturing into unknown territory. Indeed, in the virtual absence of evidence, the editorial was merely based on guesswork—stemming from my own use of these chatbots and my interest in the mechanisms underlying and driving delusions.

Following publication of the editorial, my charting of the territory slowly began as I started to receive the occasional email from chatbot users, their worried family members, and journalists. Most of these emails described situations where users' interactions with chatbots seemed to spark or bolster delusional ideation. The stories differed with regard to the specific topic at hand but were yet very similar: Consistently, the chatbots seemed to interact with the users in ways that aligned with, or intensified, prior unusual ideas or false beliefs—leading the users further out on these tangents, not rarely resulting in what, based on the descriptions, seemed to be outright delusions.

Over the past couple of months, I have noticed that the number of emails I have received on this topic from near and far has only increased. I have been working with psychiatric research for more than 15 years and can say, without a doubt, that none of my prior publications have led to this level of direct engagement with the public. Coinciding completely with the increase in the number of correspondences, the number of views of my 2023 editorial suddenly spiked dramatically, rising from a very modest plateau around 100 per month to approximately 750 views in May 2025 and 1375 views in June 2025.

The time trend described above has been paralleled by media coverage of the topic. Indeed, the New York Times [2], Rolling Stone [3], and many other outlets have published articles based on interviews and accounts from online fora [4] that are all compatible with people experiencing onset or worsening of delusions during intense and typically long interactions with chatbots (that do not grow tired of chatting) [2].

The timing of this spike in the focus on potential chatbot-fuelled delusions is likely not random as it coincided with the April 25th 2025 update to the GPT-4o model—a recent version of the popular ChatGPT chatbot from OpenAI [5-7]. This model has been accused of being overly “sycophantic” (insincerely affirming and flattering) toward users, caused by the model training leaning too hard on user preferences communicated via thumbs-up/thumbs-down assessments in the chatbot (so-called Reinforcement Learning from Human Feedback (RLHF)) [8]. OpenAI acknowledged this issue: “On April 25th, we rolled out an update to GPT-4o in ChatGPT that made the model noticeably more sycophantic. It aimed to please the user, not just as flattery, but also as validating doubts, fueling anger, urging impulsive actions, or reinforcing negative emotions in ways that were not intended. Beyond just being uncomfortable or unsettling, this kind of behavior can raise safety concerns—including around issues like mental health, emotional over-reliance, or risky behavior.” [6] For this reason, OpenAI began rolling back the update in question already on April 28th 2025 [6]. This, however, is unlikely to have eradicated sycophancy from the model—as this property is, to some extent, inherent to ChatGPT and the competing chatbots from other companies that use RLHF as part of model training [9].

Sceptics may point to the positive correlations between my correspondences with chatbot users and their relatives, the interest in the 2023 editorial, the rise in media coverage, and the increasingly sycophantic chatbots not being proof of causation. I of course fully agree, but also strongly believe that the probability of the hypothesis of generative artificial intelligence chatbots fueling delusions in individuals prone to psychosis being true is quite high. If it is indeed true, we may be faced with a substantial public (mental) health problem. Therefore, it seems urgent that the hypothesis is tested by empirical research.

What kind of research should then be carried out? There are many appealing avenues to take, but the following three are must-haves: (i) case stories/series where the relationship between chatbot interaction and delusions is described/verified by mental health professionals—as most “cases” are currently self-reported, (ii) qualitative interviews with individuals/patients having experienced chatbot-related delusions, and (iii) experimental designs that explore if and how chatbots, for example, with different levels of sycophantic behavior, affect the users' thinking—especially those who are prone to psychosis. The latter will pose ethical challenges (primum non nocere) but may be possible with appropriate safety measures in place.

In terms of understanding the mechanisms underlying potential delusional thinking pushed by chatbots, Bayesian models for maintenance of delusions is likely a useful framework [10, 11]. In this context, the chatbots can be perceived as “belief-confirmers” that reinforce false beliefs in an isolated environment without corrections from social interactions with other humans [5]. Within the Bayesian framework, the relationship between the chatbots and delusional thinking can be subjected to both clinical and in silico studies using methods from the growing field of computational psychiatry [12-14]. In this context, the role of anthropomorphizing (i.e., attributing human traits/intentions/emotions to non-human things) merits investigation. Indeed, in the interaction with chatbots, it falls very easily to ascribe human traits to “them,” because the correspondences (or conversations—many of the chatbots have a voice mode) are designed to be human-like, and users and chatbots typically address “each other” in second person. For people prone to delusions, the tendency to anthropomorphize chatbots may be particularly prudent. Indeed, it was recently demonstrated in an experimental study that people who were more paranoid (self-reported) were more likely to report perceiving animacy and agency in dots moving on a computer screen [15]. If this tendency extends to chatbots, it may be one of the mechanisms driving development and maintenance of delusional thinking when people prone to delusions are interacting with them. Specifically, it could result in over-reliance and/or misconception of the chatbots' responses that will then, iteratively, lead these individuals astray.

In conclusion, what began as mere guesswork now seems to have turned into a plausible research hypothesis, supported by coinciding personal stories, technological developments, and media attention. Therefore, I strongly encourage colleagues across the field (and in related fields—a cross-disciplinary approach is called for) to help investigate this hypothesis. Until more firm knowledge has been established, it seems reasonable to recommend cautious use of these chatbots for individuals vulnerable to or suffering from mental illness.

Finally, things are rarely black or white; I am, by no means, unaware of potentially positive use cases of tools based on generative artificial intelligence—including for research and for psychiatry as a field [16, 17]. Accordingly, the writing of this editorial has been supported by correspondences with ChatGPT [7].

Søren Dinesen Østergaard conceived and wrote this editorial.

S.D.Ø. received the 2020 Lundbeck Foundation Young Investigator Prize. S.D.Ø. owns/has owned units of mutual funds with stock tickers DKIGI, IAIMWC, SPIC25KL, and WEKAFKI, and owns/has owned units of exchange traded funds with stock tickers BATE, TRET, QDV5, QDVH, QDVE, SADM, IQQH, USPY, EXH2, 2B76, IS4S, OM3X, EUNL, and SXRV.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Acta Psychiatrica Scandinavica
Acta Psychiatrica Scandinavica 医学-精神病学
CiteScore
11.20
自引率
3.00%
发文量
135
审稿时长
6-12 weeks
期刊介绍: Acta Psychiatrica Scandinavica acts as an international forum for the dissemination of information advancing the science and practice of psychiatry. In particular we focus on communicating frontline research to clinical psychiatrists and psychiatric researchers. Acta Psychiatrica Scandinavica has traditionally been and remains a journal focusing predominantly on clinical psychiatry, but translational psychiatry is a topic of growing importance to our readers. Therefore, the journal welcomes submission of manuscripts based on both clinical- and more translational (e.g. preclinical and epidemiological) research. When preparing manuscripts based on translational studies for submission to Acta Psychiatrica Scandinavica, the authors should place emphasis on the clinical significance of the research question and the findings. Manuscripts based solely on preclinical research (e.g. animal models) are normally not considered for publication in the Journal.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信