{"title":"论计算阈值作为治理策略的局限性","authors":"Sara Hooker","doi":"arxiv-2407.05694","DOIUrl":null,"url":null,"abstract":"At face value, this essay is about understanding a fairly esoteric governance\ntool called compute thresholds. However, in order to grapple with whether these\nthresholds will achieve anything, we must first understand how they came to be.\nThis requires engaging with a decades-old debate at the heart of computer\nscience progress, namely, is bigger always better? Hence, this essay may be of\ninterest not only to policymakers and the wider public but also to computer\nscientists interested in understanding the role of compute in unlocking\nbreakthroughs. Does a certain inflection point of compute result in changes to\nthe risk profile of a model? This discussion is increasingly urgent given the\nwide adoption of governance approaches that suggest greater compute equates\nwith higher propensity for harm. Several leading frontier AI companies have\nreleased responsible scaling policies. Both the White House Executive Orders on\nAI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point\noperations as a way to identify more powerful systems. What is striking about\nthe choice of compute thresholds to-date is that no models currently deployed\nin the wild fulfill the current criteria set by the EO. This implies that the\nemphasis is often not on auditing the risks and harms incurred by currently\ndeployed models - but rather is based upon the belief that future levels of\ncompute will introduce unforeseen new risks. A key conclusion of this essay is\nthat compute thresholds as currently implemented are shortsighted and likely to\nfail to mitigate risk. Governance that is overly reliant on compute fails to\nunderstand that the relationship between compute and risk is highly uncertain\nand rapidly changing. It also overestimates our ability to predict what\nabilities emerge at different scales. This essay ends with recommendations for\na better way forward.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"2016 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"On the Limitations of Compute Thresholds as a Governance Strategy\",\"authors\":\"Sara Hooker\",\"doi\":\"arxiv-2407.05694\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"At face value, this essay is about understanding a fairly esoteric governance\\ntool called compute thresholds. However, in order to grapple with whether these\\nthresholds will achieve anything, we must first understand how they came to be.\\nThis requires engaging with a decades-old debate at the heart of computer\\nscience progress, namely, is bigger always better? Hence, this essay may be of\\ninterest not only to policymakers and the wider public but also to computer\\nscientists interested in understanding the role of compute in unlocking\\nbreakthroughs. Does a certain inflection point of compute result in changes to\\nthe risk profile of a model? This discussion is increasingly urgent given the\\nwide adoption of governance approaches that suggest greater compute equates\\nwith higher propensity for harm. Several leading frontier AI companies have\\nreleased responsible scaling policies. Both the White House Executive Orders on\\nAI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point\\noperations as a way to identify more powerful systems. What is striking about\\nthe choice of compute thresholds to-date is that no models currently deployed\\nin the wild fulfill the current criteria set by the EO. This implies that the\\nemphasis is often not on auditing the risks and harms incurred by currently\\ndeployed models - but rather is based upon the belief that future levels of\\ncompute will introduce unforeseen new risks. A key conclusion of this essay is\\nthat compute thresholds as currently implemented are shortsighted and likely to\\nfail to mitigate risk. Governance that is overly reliant on compute fails to\\nunderstand that the relationship between compute and risk is highly uncertain\\nand rapidly changing. It also overestimates our ability to predict what\\nabilities emerge at different scales. This essay ends with recommendations for\\na better way forward.\",\"PeriodicalId\":501168,\"journal\":{\"name\":\"arXiv - CS - Emerging Technologies\",\"volume\":\"2016 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Emerging Technologies\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.05694\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.05694","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
从表面上看,这篇文章的主题是了解一种名为计算阈值的相当深奥的治理工具。然而,为了探讨计算阈值是否能实现任何目标,我们必须首先了解计算阈值是如何产生的。这就需要讨论计算机科学进步的核心问题,即 "是否越大就越好?"这一长达数十年之久的争论。因此,这篇文章不仅对政策制定者和广大公众有意义,而且对有兴趣了解计算在实现突破中的作用的计算机科学家也有意义。计算的某个拐点是否会导致模型的风险状况发生变化?这一问题的讨论越来越紧迫,因为人们普遍采用的治理方法认为,计算能力越强,造成危害的可能性就越大。一些领先的前沿人工智能公司已经发布了负责任的扩展政策。白宫关于人工智能安全的行政命令(EO)和欧盟人工智能法案都将 FLOP 或浮点运算作为识别更强大系统的一种方式。迄今为止,在计算阈值的选择上令人震惊的是,目前部署在野外的模型都不符合 EO 目前设定的标准。这意味着,重点往往不在于审核当前部署的模型所带来的风险和危害--而是基于这样一种信念,即未来的计算水平将带来不可预见的新风险。本文的一个重要结论是,目前实施的计算阈值是短视的,很可能无法降低风险。过度依赖计算的治理方式未能理解计算与风险之间的关系是高度不确定和快速变化的。同时,它也高估了我们预测不同规模下可能出现的风险的能力。本文最后提出了更好的发展建议。
On the Limitations of Compute Thresholds as a Governance Strategy
At face value, this essay is about understanding a fairly esoteric governance
tool called compute thresholds. However, in order to grapple with whether these
thresholds will achieve anything, we must first understand how they came to be.
This requires engaging with a decades-old debate at the heart of computer
science progress, namely, is bigger always better? Hence, this essay may be of
interest not only to policymakers and the wider public but also to computer
scientists interested in understanding the role of compute in unlocking
breakthroughs. Does a certain inflection point of compute result in changes to
the risk profile of a model? This discussion is increasingly urgent given the
wide adoption of governance approaches that suggest greater compute equates
with higher propensity for harm. Several leading frontier AI companies have
released responsible scaling policies. Both the White House Executive Orders on
AI Safety (EO) and the EU AI Act encode the use of FLOP or floating-point
operations as a way to identify more powerful systems. What is striking about
the choice of compute thresholds to-date is that no models currently deployed
in the wild fulfill the current criteria set by the EO. This implies that the
emphasis is often not on auditing the risks and harms incurred by currently
deployed models - but rather is based upon the belief that future levels of
compute will introduce unforeseen new risks. A key conclusion of this essay is
that compute thresholds as currently implemented are shortsighted and likely to
fail to mitigate risk. Governance that is overly reliant on compute fails to
understand that the relationship between compute and risk is highly uncertain
and rapidly changing. It also overestimates our ability to predict what
abilities emerge at different scales. This essay ends with recommendations for
a better way forward.