把实验停下来!让这个AI之夏漫长些——马斯克等呼吁暂缓巨型AI训练
文库划重点:现在,我们成功地创建了强大的人工智能系统,可以在这个“人工智能之夏”中获得回报,为所有人的明确利益设计这些系统,并给社会一个适应的机会。——但实验室系统训练和公开信签名都在继续。
“加快开发强大的AI治理系统”
当全社会都在为AI的进化而欢呼雀跃时,多位科技领袖也表达了深切担忧。
据生命未来研究所官网,3月22日,生命未来研究所(Future of Life)向全社会发布了一封《暂停大型人工智能研究》的公开信,呼吁所有人工智能实验室立即暂停比GPT-4更强大的人工智能系统的训练,暂停时间至少为6个月。该机构以“引导变革性技术造福生活,远离极端的大规模风险”为使命。
信中写道“只有在我们确信它们的效果是积极的,风险是可控的情况下,才应该开发强大的人工智能系统。”马斯克此前曾多次表达对人工智能的担忧,认为人工智能是未来人类文明最大的风险之一,威胁程度远高于车祸、飞机失事、毒品泛滥等甚至比核武器都危险得多。
截至目前,马斯克、苹果联合创始人Steve Wozniak、Stability AI创始人Emad Mostaque等上千名科技大佬和AI专家已经签署公开信。
值得一提的是,OpenAI CEO奥特曼在与MIT研究科学家Lex Fridman的最新对话中指出,AI已经出现其无法解释的推理能力,同时承认“AI杀死人类”有一定可能性。
未来生命研究所的这封公开信题为“暂停巨型AI实验:一封公开信”,于29日解封。
以下是公开信中文翻译:
暂停巨型AI实验:一封公开信
我们呼吁所有AI实验室立即暂停训练比GPT-4更强大的AI系统至少6个月。
正如广泛的研究和顶级人工智能实验室所承认的,人工智能系统对社会和人类构成较大的风险。《阿西洛马人工智能原则》中指出,高级人工智能可能代表着地球上生命历史的深刻变革,应该以相应的谨慎和资源进行规划和管理。然而,尽管最近几个月人工智能实验室掀起AI狂潮,开发和部署越来越强大的数字大脑,但目前没有人能理解、预测或可靠地控制AI系统,也没有相应水平的规划和管理。
现在,人工智能在一般任务上变得与人类有竞争力,我们必须问自己:我们应该让机器在信息渠道中宣传不真实的信息吗?我们是否应该把所有的工作都自动化,包括那些有成就感的工作?我们是否应该开发非人类的大脑,使其最终超过人类数量,胜过人类的智慧,淘汰并取代人类?我们是否应该冒着失去对我们文明控制的风险?这样的决定绝不能委托给未经选举的技术领袖。只有当我们确信强大的人工智能系统的效果是积极的,其风险是可控的,才应该开发。同时,这种信心必须得到验证,并随着系统的潜在影响的大小而加强。OpenAI最近关于人工智能的声明中指出,在开始训练未来的系统之前,可能必须得到独立的审查,对于最先进的努力,同意限制用于创建新模型的计算增长速度。我们同意,现在就该采取行动。
因此,我们呼吁所有人工智能实验室立即暂停比GPT-4更强大的人工智能系统的训练,时间至少持续6个月。这种暂停应该是公开的、可核查的,并包括所有关键参与者者。如果这种禁令不能迅速实施,政府应该介入并制定暂停令。
人工智能实验室和独立专家应在暂停期间,共同制定和实施一套先进的人工智能设计和开发的共享安全协议,由独立的外部专家进行严格的审查和监督。这些协议应该确保遵守协议的系统是安全的。值得一提的是,这并不意味着暂停一般的人工智能开发,只是从危险的竞赛中退后一步,限制不可预测的研究和开发。
人工智能的研究和开发应该重新聚焦于,使目前最先进和强大的系统更加准确、安全、可解释、透明、稳健、一致、值得信赖和忠诚。
同时,人工智能开发者必须与政策制定者合作,大幅加快开发强大的人工智能治理系统。这些至少应该包括:专门针对人工智能的监管机构;监督和跟踪高能力的人工智能系统和大型计算能力的硬件;出处和水印系统帮助区分真实和合成,并跟踪模型泄漏;强大的审查和认证生态系统;对人工智能造成的伤害承担责任;为人工智能安全技术研究提供强大的公共资金以及资源充足的机构,以应对人工智能可能导致的巨大的经济和政治破坏。
人类可以通过人工智能享受繁荣的未来。现在,我们成功地创建了强大的人工智能系统,可以在这个“人工智能之夏”中获得回报,为所有人的明确利益设计这些系统,并给社会一个适应的机会。停止使用其他技术可能会对社会造成灾难性影响,因此我们必须保持准备状态。让我们享受一个漫长的人工智能夏天,而不是匆忙进入秋天。
截至发稿,这封信已有1127名科技领袖和研究人员签名,除了马斯克、辛顿和马库斯之外,还包括图灵奖得主约书亚·本希奥(Yoshua Bengio)、《人工智能:现代方法》作者斯图尔特·罗素(Stuart Russell)、苹果公司联合创始人史蒂夫·沃兹尼亚克(Steve Wozniak)、Stability AI首席执行官埃马德·莫斯塔克(Emad Mostaque)等科技界领袖人物。
以下是公开信原文英文:
Pause Giant AI Experiments: An Open Letter
We call on all AI labs to immediately pause for at least 6 months the training of AI systems more powerful than GPT-4.
AI systems with human-competitive intelligence can pose profound risks to society and humanity, as shown by extensive research[1] and acknowledged by top AI labs.[2] As stated in the widely-endorsed Asilomar AI Principles, Advanced AI could represent a profound change in the history of life on Earth, and should be planned for and managed with commensurate care and resources. Unfortunately, this level of planning and management is not happening, even though recent months have seen AI labs locked in an out-of-control race to develop and deploy ever more powerful digital minds that no one – not even their creators – can understand, predict, or reliably control.
Contemporary AI systems are now becoming human-competitive at general tasks,[3] and we must ask ourselves: Should we let machines flood our information channels with propaganda and untruth? Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization? Such decisions must not be delegated to unelected tech leaders. Powerful AI systems should be developed only once we are confident that their effects will be positive and their risks will be manageable. This confidence must be well justified and increase with the magnitude of a system's potential effects. OpenAI's recent statement regarding artificial general intelligence, states that "At some point, it may be important to get independent review before starting to train future systems, and for the most advanced efforts to agree to limit the rate of growth of compute used for creating new models." We agree. That point is now.
Therefore, we call on all AI labs to immediately pause for at least 6 months the training of AI systems more powerful than GPT-4. This pause should be public and verifiable, and include all key actors. If such a pause cannot be enacted quickly, governments should step in and institute a moratorium.
AI labs and independent experts should use this pause to jointly develop and implement a set of shared safety protocols for advanced AI design and development that are rigorously audited and overseen by independent outside experts. These protocols should ensure that systems adhering to them are safe beyond a reasonable doubt.[4] This does not mean a pause on AI development in general, merely a stepping back from the dangerous race to ever-larger unpredictable black-box models with emergent capabilities.
AI research and development should be refocused on making today's powerful, state-of-the-art systems more accurate, safe, interpretable, transparent, robust, aligned, trustworthy, and loyal.
In parallel, AI developers must work with policymakers to dramatically accelerate development of robust AI governance systems. These should at a minimum include: new and capable regulatory authorities dedicated to AI; oversight and tracking of highly capable AI systems and large pools of computational capability; provenance and watermarking systems to help distinguish real from synthetic and to track model leaks; a robust auditing and certification ecosystem; liability for AI-caused harm; robust public funding for technical AI safety research; and well-resourced institutions for coping with the dramatic economic and political disruptions (especially to democracy) that AI will cause.
Humanity can enjoy a flourishing future with AI. Having succeeded in creating powerful AI systems, we can now enjoy an "AI summer" in which we reap the rewards, engineer these systems for the clear benefit of all, and give society a chance to adapt. Society has hit pause on other technologies with potentially catastrophic effects on society.[5] We can do so here. Let's enjoy a long AI summer, not rush unprepared into a fall.
以下链接为生命未来研究所官网公开信网页、全信注释及签名处(任何人都可以签名):
https://futureoflife.org/open-letter/pause-giant-ai-experiments/
(正文完)
(中文原文标题:《“把实验停下来!”马斯克等超1000人签署公开信:所有实验室,立即暂停训练比GPT-4更强大的AI》。收录有节选并经哈希力量二次综合整理。)
(中文收录源链接:https://www.nbd.com.cn/articles/2023-03-29/2733133.html)
收录于哈希力量,手机站省略本文固定网址