论计算阈值作为治理策略的局限性

Sara Hooker
{"title":"论计算阈值作为治理策略的局限性","authors":"Sara Hooker","doi":"arxiv-2407.05694","DOIUrl":null,"url":null,"abstract":"At face value, this essay is about understanding a fairly esoteric governance\ntool called compute thresholds. However, in order to grapple with whether these\nthresholds will achieve anything, we must first understand how they came to be.\nThis requires engaging with a decades-old debate at the heart of computer\nscience progress, namely, is bigger always better? Hence, this essay may be of\ninterest not only to policymakers and the wider public but also to computer\nscientists interested in understanding the role of compute in unlocking\nbreakthroughs. Does a certain inflection point of compute result in changes to\nthe risk profile of a model? This discussion is increasingly urgent given the\nwide adoption of governance approaches that suggest greater compute equates\nwith higher propensity for harm. Several leading frontier AI companies have\nreleased responsible scaling policies. Both the White House Executive Orders on\nAI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point\noperations as a way to identify more powerful systems. What is striking about\nthe choice of compute thresholds to-date is that no models currently deployed\nin the wild fulfill the current criteria set by the EO. This implies that the\nemphasis is often not on auditing the risks and harms incurred by currently\ndeployed models - but rather is based upon the belief that future levels of\ncompute will introduce unforeseen new risks. A key conclusion of this essay is\nthat compute thresholds as currently implemented are shortsighted and likely to\nfail to mitigate risk. Governance that is overly reliant on compute fails to\nunderstand that the relationship between compute and risk is highly uncertain\nand rapidly changing. It also overestimates our ability to predict what\nabilities emerge at different scales. This essay ends with recommendations for\na better way forward.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"2016 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"On the Limitations of Compute Thresholds as a Governance Strategy\",\"authors\":\"Sara Hooker\",\"doi\":\"arxiv-2407.05694\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"At face value, this essay is about understanding a fairly esoteric governance\\ntool called compute thresholds. However, in order to grapple with whether these\\nthresholds will achieve anything, we must first understand how they came to be.\\nThis requires engaging with a decades-old debate at the heart of computer\\nscience progress, namely, is bigger always better? Hence, this essay may be of\\ninterest not only to policymakers and the wider public but also to computer\\nscientists interested in understanding the role of compute in unlocking\\nbreakthroughs. Does a certain inflection point of compute result in changes to\\nthe risk profile of a model? This discussion is increasingly urgent given the\\nwide adoption of governance approaches that suggest greater compute equates\\nwith higher propensity for harm. Several leading frontier AI companies have\\nreleased responsible scaling policies. Both the White House Executive Orders on\\nAI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point\\noperations as a way to identify more powerful systems. What is striking about\\nthe choice of compute thresholds to-date is that no models currently deployed\\nin the wild fulfill the current criteria set by the EO. This implies that the\\nemphasis is often not on auditing the risks and harms incurred by currently\\ndeployed models - but rather is based upon the belief that future levels of\\ncompute will introduce unforeseen new risks. A key conclusion of this essay is\\nthat compute thresholds as currently implemented are shortsighted and likely to\\nfail to mitigate risk. Governance that is overly reliant on compute fails to\\nunderstand that the relationship between compute and risk is highly uncertain\\nand rapidly changing. It also overestimates our ability to predict what\\nabilities emerge at different scales. This essay ends with recommendations for\\na better way forward.\",\"PeriodicalId\":501168,\"journal\":{\"name\":\"arXiv - CS - Emerging Technologies\",\"volume\":\"2016 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Emerging Technologies\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.05694\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.05694","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

从表面上看,这篇文章的主题是了解一种名为计算阈值的相当深奥的治理工具。然而,为了探讨计算阈值是否能实现任何目标,我们必须首先了解计算阈值是如何产生的。这就需要讨论计算机科学进步的核心问题,即 "是否越大就越好?"这一长达数十年之久的争论。因此,这篇文章不仅对政策制定者和广大公众有意义,而且对有兴趣了解计算在实现突破中的作用的计算机科学家也有意义。计算的某个拐点是否会导致模型的风险状况发生变化?这一问题的讨论越来越紧迫,因为人们普遍采用的治理方法认为,计算能力越强,造成危害的可能性就越大。一些领先的前沿人工智能公司已经发布了负责任的扩展政策。白宫关于人工智能安全的行政命令(EO)和欧盟人工智能法案都将 FLOP 或浮点运算作为识别更强大系统的一种方式。迄今为止,在计算阈值的选择上令人震惊的是,目前部署在野外的模型都不符合 EO 目前设定的标准。这意味着,重点往往不在于审核当前部署的模型所带来的风险和危害--而是基于这样一种信念,即未来的计算水平将带来不可预见的新风险。本文的一个重要结论是,目前实施的计算阈值是短视的,很可能无法降低风险。过度依赖计算的治理方式未能理解计算与风险之间的关系是高度不确定和快速变化的。同时,它也高估了我们预测不同规模下可能出现的风险的能力。本文最后提出了更好的发展建议。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
On the Limitations of Compute Thresholds as a Governance Strategy
At face value, this essay is about understanding a fairly esoteric governance tool called compute thresholds. However, in order to grapple with whether these thresholds will achieve anything, we must first understand how they came to be. This requires engaging with a decades-old debate at the heart of computer science progress, namely, is bigger always better? Hence, this essay may be of interest not only to policymakers and the wider public but also to computer scientists interested in understanding the role of compute in unlocking breakthroughs. Does a certain inflection point of compute result in changes to the risk profile of a model? This discussion is increasingly urgent given the wide adoption of governance approaches that suggest greater compute equates with higher propensity for harm. Several leading frontier AI companies have released responsible scaling policies. Both the White House Executive Orders on AI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point operations as a way to identify more powerful systems. What is striking about the choice of compute thresholds to-date is that no models currently deployed in the wild fulfill the current criteria set by the EO. This implies that the emphasis is often not on auditing the risks and harms incurred by currently deployed models - but rather is based upon the belief that future levels of compute will introduce unforeseen new risks. A key conclusion of this essay is that compute thresholds as currently implemented are shortsighted and likely to fail to mitigate risk. Governance that is overly reliant on compute fails to understand that the relationship between compute and risk is highly uncertain and rapidly changing. It also overestimates our ability to predict what abilities emerge at different scales. This essay ends with recommendations for a better way forward.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信