Evidence of interrelated cognitive-like capabilities in large language models: Indications of artificial general intelligence or achievement?

IF 4.3 3区 材料科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
David Ilić , Gilles E. Gignac
{"title":"Evidence of interrelated cognitive-like capabilities in large language models: Indications of artificial general intelligence or achievement?","authors":"David Ilić ,&nbsp;Gilles E. Gignac","doi":"10.1016/j.intell.2024.101858","DOIUrl":null,"url":null,"abstract":"<div><p>Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (<em>Gf</em>), domain-specific knowledge (<em>Gkn</em>), reading/writing (<em>Grw</em>), and quantitative knowledge (<em>Gq</em>), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined <em>Gkn</em>/<em>Grw</em> group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and <em>Gkn</em>/<em>Grw</em> factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.</p></div>","PeriodicalId":3,"journal":{"name":"ACS Applied Electronic Materials","volume":null,"pages":null},"PeriodicalIF":4.3000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0160289624000527/pdfft?md5=fca3c71c01b2f51c86dae15548627371&pid=1-s2.0-S0160289624000527-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Applied Electronic Materials","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0160289624000527","RegionNum":3,"RegionCategory":"材料科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (Gf), domain-specific knowledge (Gkn), reading/writing (Grw), and quantitative knowledge (Gq), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined Gkn/Grw group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and Gkn/Grw factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.

大型语言模型中相互关联的类认知能力的证据:人工通用智能的迹象还是成就?
大型语言模型(LLM)是一种先进的人工智能(AI)系统,可以完成人类智力测试中常见的各种任务,如定义词语、进行计算和语言推理。LLM 的能力也存在很大的个体差异。鉴于在人类样本中持续观察到正的流形和一般智能因子,以及群体级因子(如结晶智能),我们假设 LLM 测试分数也可能表现出正的相互关系,这有可能产生人工一般能力(AGA)因子和一个或多个群体级因子。基于 591 名法学硕士的样本和 12 项测试的分数,我们发现了能力的正流形和一般因子的强有力的经验证据,这些测试包括流体推理(Gf)、特定领域知识(Gkn)、阅读/写作(Grw)和定量知识(Gq)。此外,我们还发现了一个 Gkn/Grw 组级综合因子。最后,LLM参数的数量与一般能力因子和Gkn/Grw因子得分呈正相关,尽管其效果呈递减趋势。我们对结果的解释是,LLM 与人类的认知能力一样,可能在处理信息和解决问题方面具有共同的潜在效率,但 LLM 是否主要表现为成就/专长而非智力仍有待确定。最后,虽然参数数量较多的模型表现出较强的类似认知能力,类似于神经元密度较高与人类一般智力之间的联系,但其他特征也必须参与其中。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.20
自引率
4.30%
发文量
567
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信