Infusing Advanced AGIs with Human-Like Value Systems

B. Goertzel
{"title":"Infusing Advanced AGIs with Human-Like Value Systems","authors":"B. Goertzel","doi":"10.55613/jeet.v26i1.51","DOIUrl":null,"url":null,"abstract":"Two theses are proposed, regarding the future evolution of the value systems of advanced AGI systems. The Value Learning Thesis is a semi-formalized version of the idea that, if an AGI system is taught human values in an interactive and experiential way as its intelligence increases toward human level, it will likely adopt these human values in a genuine way. The Value Evolution Thesis is a semi-formalized version of the idea that if an AGI system begins with human-like values, and then iteratively modifies itself, it will end up in roughly the same future states as a population of human beings engaged with progressively increasing their own intelligence (e.g. by cyborgification or brain modification). Taken together, these theses suggest a worldview in which raising young AGIs to have human-like values is a sensible thing to do, and likely to produce a future that is generally desirable in a human sense. \nWhile these two theses are far from definitively proven, I argue that they are more solid and more relevant to the actual future of AGI than Bostrom’s “Instrumental Convergence Thesis” and “Orthogonality Thesis” which are core to the basis of his argument (in his book Superintelligence) for fearing ongoing AGI development and placing AGI R&D under strict governmental control. In the context of fleshing out this argument, previous publications and discussions by Richard Loosemore and Kaj Sotala are discussed in some detail.","PeriodicalId":157018,"journal":{"name":"Journal of Ethics and Emerging Technologies","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"14","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Ethics and Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.55613/jeet.v26i1.51","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 14

Abstract

Two theses are proposed, regarding the future evolution of the value systems of advanced AGI systems. The Value Learning Thesis is a semi-formalized version of the idea that, if an AGI system is taught human values in an interactive and experiential way as its intelligence increases toward human level, it will likely adopt these human values in a genuine way. The Value Evolution Thesis is a semi-formalized version of the idea that if an AGI system begins with human-like values, and then iteratively modifies itself, it will end up in roughly the same future states as a population of human beings engaged with progressively increasing their own intelligence (e.g. by cyborgification or brain modification). Taken together, these theses suggest a worldview in which raising young AGIs to have human-like values is a sensible thing to do, and likely to produce a future that is generally desirable in a human sense. While these two theses are far from definitively proven, I argue that they are more solid and more relevant to the actual future of AGI than Bostrom’s “Instrumental Convergence Thesis” and “Orthogonality Thesis” which are core to the basis of his argument (in his book Superintelligence) for fearing ongoing AGI development and placing AGI R&D under strict governmental control. In the context of fleshing out this argument, previous publications and discussions by Richard Loosemore and Kaj Sotala are discussed in some detail.
为先进的人工智能系统注入类似人类的价值体系
针对先进AGI系统价值系统的未来发展,提出了两篇论文。价值学习论文是一种半形式化的观点,即如果一个人工智能系统在其智能向人类水平增长的过程中,以一种互动和体验的方式教授人类价值观,它可能会以一种真正的方式接受这些人类价值观。价值进化理论是一种半形式化的观点,即如果一个AGI系统从类似人类的价值观开始,然后迭代地修改自己,那么它最终将处于与人类群体大致相同的未来状态,并逐渐增加自己的智力(例如,通过电子化或大脑改造)。综上所述,这些论点提出了一种世界观,在这种世界观中,培养年轻的agi具有类似人类的价值观是一件明智的事情,并且可能产生一个在人类意义上普遍可取的未来。虽然这两个论点远未得到明确证明,但我认为它们比博斯特罗姆的“工具收敛论点”和“正交论点”更可靠,与AGI的实际未来更相关,这是他(在他的书《超级智能》中)担心持续的AGI发展并将AGI研发置于严格的政府控制之下的论点的核心基础。在充实这一论点的背景下,对Richard Loosemore和Kaj Sotala以前的出版物和讨论进行了一些详细的讨论。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信