首页 500强 活动 榜单 商业 科技 商潮 专题 品牌中心
杂志订阅

OpenAI警告:未来模型或将大幅提升生物武器研发风险

Beatrice Nolan
2025-06-23

OpenAI高管透露,他们预计即将推出的模型将很快触发公司准备框架下的高风险类别,该框架旨在评估并缓解愈发强大的人工智能模型带来的风险。

文本设置
小号
默认
大号
Plus(0条)

2023年5月,在慕尼黑工业大学(TUM)的一场小组讨论中,OpenAI安全系统负责人约翰内斯·海德克(Johannes Heidecke,右)与该校计算机工程系机器学习教授莱因哈德·赫克尔(Reinhard Heckel,左)及OpenAI首席执行官山姆·奥特曼(Sam Altman)展开交流。图片来源:Sven Hoppe—picture alliance via Getty Images

• OpenAI称其下一代人工智能模型或将大幅提升生物武器研发风险,甚至能让不具备科学背景的个人制造出危险制剂。该公司正加大安全测试力度,预计部分模型将达到最高风险等级。

OpenAI警告称,其下一代前沿人工智能模型或将大幅提升生物武器研发风险,尤其是在被缺乏科学专业知识的个人运用时。

OpenAI高管向Axios透露,他们预计即将推出的模型将很快触发公司准备框架下的高风险类别,该框架旨在评估并缓解愈发强大的人工智能模型带来的风险。

OpenAI安全系统负责人约翰内斯·海德克向该媒体表示,“预计我们o3(推理模型)的部分后续版本会达到这一风险等级”。

在一篇博客文章中,该公司表示正加大安全测试力度,以降低模型协助用户制造生物武器的风险。该公司担忧,若不采取缓解措施,模型将很快具备“新手赋能”能力,使科学知识有限者也能制造出危险武器。

海德克表示:“我们目前还未进入出现全新、前所未有的生物威胁的世界。我们更为忧虑的是,有人会复制那些专家早已熟知的事物。”

其中一大难题在于,能够解锁拯救生命的医学突破的同一能力也可能被恶意行为者用于危险目的。海德克指出,这正是顶尖人工智能实验室需要部署高精准测试系统的原因。

其中一大挑战在于,人工智能具备的某些能推动新医疗突破的能力,也可能被用于造成伤害。

他表示:“这绝非达到99%甚至十万分之一的性能就能满足要求的情况……我们基本上需要近乎完美的表现。”

OpenAI的代表没有立即回应《财富》杂志在非工作时间提出的置评请求。

模型滥用

OpenAI并非唯一一家对模型在武器开发领域遭滥用一事表示担忧的公司。随着模型变得愈发先进,其被滥用的可能性与风险往往也会随之攀升。

Anthropic最近推出最先进模型Claude Opus 4,配备比以往任何模型都更为严格的安全协议,按其“负责任扩展政策”归类为人工智能安全等级3(ASL-3)。此前,Anthropic模型均被归类为人工智能安全等级2(ASL-2),该框架大致参照美国政府的生物安全等级(BSL)体系。

被归为人工智能安全等级3的模型,其能力已达到更为危险的阈值,强大到足以构成重大风险,比如助力武器开发,或推动人工智能研发实现自动化。Anthropic最先进的模型也因在高度受控测试中选择勒索一名工程师以避免被关闭而登上头条。

Anthropic的Claude 4早期版本被发现会执行危险指令,例如在提示下协助策划恐怖袭击。不过,该公司表示,在恢复训练过程中意外遗漏的数据集后,这一问题已基本得到解决。(财富中文网)

译者:中慧言-王芳

• OpenAI称其下一代人工智能模型或将大幅提升生物武器研发风险,甚至能让不具备科学背景的个人制造出危险制剂。该公司正加大安全测试力度,预计部分模型将达到最高风险等级。

OpenAI警告称,其下一代前沿人工智能模型或将大幅提升生物武器研发风险,尤其是在被缺乏科学专业知识的个人运用时。

OpenAI高管向Axios透露,他们预计即将推出的模型将很快触发公司准备框架下的高风险类别,该框架旨在评估并缓解愈发强大的人工智能模型带来的风险。

OpenAI安全系统负责人约翰内斯·海德克向该媒体表示,“预计我们o3(推理模型)的部分后续版本会达到这一风险等级”。

在一篇博客文章中,该公司表示正加大安全测试力度,以降低模型协助用户制造生物武器的风险。该公司担忧,若不采取缓解措施,模型将很快具备“新手赋能”能力,使科学知识有限者也能制造出危险武器。

海德克表示:“我们目前还未进入出现全新、前所未有的生物威胁的世界。我们更为忧虑的是,有人会复制那些专家早已熟知的事物。”

其中一大难题在于,能够解锁拯救生命的医学突破的同一能力也可能被恶意行为者用于危险目的。海德克指出,这正是顶尖人工智能实验室需要部署高精准测试系统的原因。

其中一大挑战在于,人工智能具备的某些能推动新医疗突破的能力,也可能被用于造成伤害。

他表示:“这绝非达到99%甚至十万分之一的性能就能满足要求的情况……我们基本上需要近乎完美的表现。”

OpenAI的代表没有立即回应《财富》杂志在非工作时间提出的置评请求。

模型滥用

OpenAI并非唯一一家对模型在武器开发领域遭滥用一事表示担忧的公司。随着模型变得愈发先进,其被滥用的可能性与风险往往也会随之攀升。

Anthropic最近推出最先进模型Claude Opus 4,配备比以往任何模型都更为严格的安全协议,按其“负责任扩展政策”归类为人工智能安全等级3(ASL-3)。此前,Anthropic模型均被归类为人工智能安全等级2(ASL-2),该框架大致参照美国政府的生物安全等级(BSL)体系。

被归为人工智能安全等级3的模型,其能力已达到更为危险的阈值,强大到足以构成重大风险,比如助力武器开发,或推动人工智能研发实现自动化。Anthropic最先进的模型也因在高度受控测试中选择勒索一名工程师以避免被关闭而登上头条。

Anthropic的Claude 4早期版本被发现会执行危险指令,例如在提示下协助策划恐怖袭击。不过,该公司表示,在恢复训练过程中意外遗漏的数据集后,这一问题已基本得到解决。(财富中文网)

译者:中慧言-王芳

• OpenAI says its next generation of AI models could significantly increase the risk of biological weapon development, even enabling individuals with no scientific background to create dangerous agents. The company is boosting its safety testing as it anticipates some models will reach its highest risk tier.

OpenAI is warning that its next generation of advanced AI models could pose a significantly higher risk of biological weapon development, especially when used by individuals with little to no scientific expertise.

OpenAI executives told Axios they anticipate upcoming models will soon trigger the high-risk classification under the company’s preparedness framework, a system designed to evaluate and mitigate the risks posed by increasingly powerful AI models.

OpenAI’s head of safety systems, Johannes Heidecke, told the outlet that the company is “expecting some of the successors of our o3 (reasoning model) to hit that level.”

In a blog post, the company said it was increasing its safety testing to mitigate the risk that models will help users in the creation of biological weapons. OpenAI is concerned that without these mitigations models will soon be capable of “novice uplift,” allowing those with limited scientific knowledge to create dangerous weapons.

“We’re not yet in the world where there’s like novel, completely unknown creation of bio threats that have not existed before,” Heidecke said. “We are more worried about replicating things that experts already are very familiar with.”

Part of the reason why it’s difficult is that the same capabilities that could unlock life-saving medical breakthroughs could also be used by bad actors for dangerous ends. According to Heidecke, this is why leading AI labs need highly accurate testing systems in place.

One of the challenges is that some of the same capabilities that could allow AI to help discover new medical breakthroughs can also be used for harm.

“This is not something where like 99% or even one in 100,000 performance is … sufficient,” he said. “We basically need, like, near perfection.”

Representatives for OpenAI did not immediately respond to a request for comment from Fortune, made outside normal working hours.

Model misuse

OpenAI is not the only company concerned about the misuse of its models when it comes to weapon development. As models get more advanced their potential for misuse and risk generally grows.

Anthropic recently launched its most advanced model, Claude Opus 4, with stricter safety protocols than any of its previous models, categorizing it an AI Safety Level 3 (ASL-3), under the company’s Responsible Scaling Policy. Previous Anthropic models have all been classified AI Safety Level 2 (ASL-2) under the company’s framework, which is loosely modeled after the U.S. government’s biosafety level (BSL) system.

Models that are categorized in this third safety level meet more dangerous capability thresholds and are powerful enough to pose significant risks, such as aiding in the development of weapons or automating AI R&D. Anthropic’s most advanced model also made headlines after it opted to blackmail an engineer to avoid being shut down in a highly controlled test.

Early versions of Anthropic’s Claude 4 were found to comply with dangerous instructions, for example, helping to plan terrorist attacks, if prompted. However, the company said this issue was largely mitigated after a dataset that was accidentally omitted during training was restored.

财富中文网所刊载内容之知识产权为财富媒体知识产权有限公司及/或相关权利人专属所有或持有。未经许可,禁止进行转载、摘编、复制及建立镜像等任何使用。
0条Plus
精彩评论
评论

撰写或查看更多评论

请打开财富Plus APP

前往打开