Large language models outperform humans in identifying neuromyths but show sycophantic behavior in applied contexts

IF 3.4 Q2 NEUROSCIENCES
Eileen Richter , Markus Wolfgang Hermann Spitzer , Annabelle Morgan , Luisa Frede , Joshua Weidlich , Korbinian Moeller
{"title":"Large language models outperform humans in identifying neuromyths but show sycophantic behavior in applied contexts","authors":"Eileen Richter ,&nbsp;Markus Wolfgang Hermann Spitzer ,&nbsp;Annabelle Morgan ,&nbsp;Luisa Frede ,&nbsp;Joshua Weidlich ,&nbsp;Korbinian Moeller","doi":"10.1016/j.tine.2025.100255","DOIUrl":null,"url":null,"abstract":"<div><div><strong>Background</strong>: Neuromyths are widespread among educators, which raises concerns about misconceptions regarding the (neural) principles underlying learning in the educator population. With the increasing use of large language models (LLMs) in education, educators are increasingly relying on these for lesson planning and professional development. Therefore, if LLMs correctly identify neuromyths, they may help to dispute related misconceptions.</div><div><strong>Method</strong>: We evaluated whether LLMs can correctly identify neuromyths and whether they may hint educators to neuromyths in applied contexts when users ask questions comprising related misconceptions. Additionally, we examined whether explicitly prompting LLMs to base their answer on scientific evidence or to correct unsupported assumptions would decrease errors in identifying neuromyths.</div><div><strong>Results</strong>: LLMs outperformed humans in identifying neuromyth statements as used in previous studies. However, when presented with applied user-like questions comprising misconceptions, they struggled to highlight or dispute these. Interestingly, explicitly asking LLMs to correct unsupported assumptions increased the likelihood that misconceptions were flagged considerably, while prompting the models to rely on scientific evidence had only little effects.</div><div><strong>Conclusion</strong>: While LLMs outperformed humans at identifying isolated neuromyth statements, they struggled to hint users towards the same misconception when they were included in more applied user-like questions—presumably due to LLMs’ tendency toward sycophantic responses. This limitation suggests that, despite their potential, LLMs are not yet a reliable safeguard against the spread of neuromyths in educational settings. However, when users explicitly prompt LLMs to correct unsupported assumptions—an approach that may initially seem counterintuitive–this effectively reduced sycophantic responses.</div></div>","PeriodicalId":46228,"journal":{"name":"Trends in Neuroscience and Education","volume":"39 ","pages":"Article 100255"},"PeriodicalIF":3.4000,"publicationDate":"2025-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Trends in Neuroscience and Education","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2211949325000092","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Neuromyths are widespread among educators, which raises concerns about misconceptions regarding the (neural) principles underlying learning in the educator population. With the increasing use of large language models (LLMs) in education, educators are increasingly relying on these for lesson planning and professional development. Therefore, if LLMs correctly identify neuromyths, they may help to dispute related misconceptions.
Method: We evaluated whether LLMs can correctly identify neuromyths and whether they may hint educators to neuromyths in applied contexts when users ask questions comprising related misconceptions. Additionally, we examined whether explicitly prompting LLMs to base their answer on scientific evidence or to correct unsupported assumptions would decrease errors in identifying neuromyths.
Results: LLMs outperformed humans in identifying neuromyth statements as used in previous studies. However, when presented with applied user-like questions comprising misconceptions, they struggled to highlight or dispute these. Interestingly, explicitly asking LLMs to correct unsupported assumptions increased the likelihood that misconceptions were flagged considerably, while prompting the models to rely on scientific evidence had only little effects.
Conclusion: While LLMs outperformed humans at identifying isolated neuromyth statements, they struggled to hint users towards the same misconception when they were included in more applied user-like questions—presumably due to LLMs’ tendency toward sycophantic responses. This limitation suggests that, despite their potential, LLMs are not yet a reliable safeguard against the spread of neuromyths in educational settings. However, when users explicitly prompt LLMs to correct unsupported assumptions—an approach that may initially seem counterintuitive–this effectively reduced sycophantic responses.
大型语言模型在识别神经神话方面优于人类,但在应用环境中表现出谄媚的行为
背景:神经神话在教育工作者中普遍存在,这引起了人们对教育工作者群体中关于学习基础(神经)原理的误解的关注。随着在教育中越来越多地使用大型语言模型(llm),教育工作者越来越依赖于这些课程计划和专业发展。因此,如果法学硕士正确识别神经神话,他们可能有助于争议相关的误解。方法:我们评估法学硕士是否可以正确识别神经神话,以及当用户提出包含相关误解的问题时,他们是否可以在应用环境中提示教育者神经神话。此外,我们还研究了明确提示法学硕士基于科学证据或纠正不受支持的假设是否会减少识别神经神话的错误。结果:法学硕士在识别先前研究中使用的神经神话陈述方面优于人类。然而,当出现包含误解的应用用户类问题时,他们很难突出或反驳这些问题。有趣的是,明确要求法学硕士纠正不受支持的假设,大大增加了错误观念被标记出来的可能性,而促使模型依赖科学证据的效果却很小。结论:虽然llm在识别孤立的神经神话语句方面优于人类,但当它们被包含在更实用的用户类问题中时,它们很难暗示用户同样的误解——可能是由于llm倾向于阿谀奉承的反应。这一限制表明,尽管法学硕士有潜力,但它还不是防止神经神话在教育环境中传播的可靠保障。然而,当用户明确提示法学硕士纠正不受支持的假设时(这种方法最初可能看起来违反直觉),这有效地减少了谄媚的反应。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
6.30
自引率
6.10%
发文量
22
审稿时长
65 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信