
? OpenAI稱其下一代人工智能模型或將大幅提升生物武器研發風險,甚至能讓不具備科學背景的個人制造出危險制劑。該公司正加大安全測試力度,預計部分模型將達到最高風險等級。
OpenAI警告稱,其下一代前沿人工智能模型或將大幅提升生物武器研發風險,尤其是在被缺乏科學專業知識的個人運用時。
OpenAI高管向Axios透露,他們預計即將推出的模型將很快觸發公司準備框架下的高風險類別,該框架旨在評估并緩解愈發強大的人工智能模型帶來的風險。
OpenAI安全系統負責人約翰內斯·海德克向該媒體表示,“預計我們o3(推理模型)的部分后續版本會達到這一風險等級”。
在一篇博客文章中,該公司表示正加大安全測試力度,以降低模型協助用戶制造生物武器的風險。該公司擔憂,若不采取緩解措施,模型將很快具備“新手賦能”能力,使科學知識有限者也能制造出危險武器。
海德克表示:“我們目前還未進入出現全新、前所未有的生物威脅的世界。我們更為憂慮的是,有人會復制那些專家早已熟知的事物。”
其中一大難題在于,能夠解鎖拯救生命的醫學突破的同一能力也可能被惡意行為者用于危險目的。海德克指出,這正是頂尖人工智能實驗室需要部署高精準測試系統的原因。
其中一大挑戰在于,人工智能具備的某些能推動新醫療突破的能力,也可能被用于造成傷害。
他表示:“這絕非達到99%甚至十萬分之一的性能就能滿足要求的情況……我們基本上需要近乎完美的表現。”
OpenAI的代表沒有立即回應《財富》雜志在非工作時間提出的置評請求。
模型濫用
OpenAI并非唯一一家對模型在武器開發領域遭濫用一事表示擔憂的公司。隨著模型變得愈發先進,其被濫用的可能性與風險往往也會隨之攀升。
Anthropic最近推出最先進模型Claude Opus 4,配備比以往任何模型都更為嚴格的安全協議,按其“負責任擴展政策”歸類為人工智能安全等級3(ASL-3)。此前,Anthropic模型均被歸類為人工智能安全等級2(ASL-2),該框架大致參照美國政府的生物安全等級(BSL)體系。
被歸為人工智能安全等級3的模型,其能力已達到更為危險的閾值,強大到足以構成重大風險,比如助力武器開發,或推動人工智能研發實現自動化。Anthropic最先進的模型也因在高度受控測試中選擇勒索一名工程師以避免被關閉而登上頭條。
Anthropic的Claude 4早期版本被發現會執行危險指令,例如在提示下協助策劃恐怖襲擊。不過,該公司表示,在恢復訓練過程中意外遺漏的數據集后,這一問題已基本得到解決。(財富中文網)
譯者:中慧言-王芳
? OpenAI稱其下一代人工智能模型或將大幅提升生物武器研發風險,甚至能讓不具備科學背景的個人制造出危險制劑。該公司正加大安全測試力度,預計部分模型將達到最高風險等級。
OpenAI警告稱,其下一代前沿人工智能模型或將大幅提升生物武器研發風險,尤其是在被缺乏科學專業知識的個人運用時。
OpenAI高管向Axios透露,他們預計即將推出的模型將很快觸發公司準備框架下的高風險類別,該框架旨在評估并緩解愈發強大的人工智能模型帶來的風險。
OpenAI安全系統負責人約翰內斯·海德克向該媒體表示,“預計我們o3(推理模型)的部分后續版本會達到這一風險等級”。
在一篇博客文章中,該公司表示正加大安全測試力度,以降低模型協助用戶制造生物武器的風險。該公司擔憂,若不采取緩解措施,模型將很快具備“新手賦能”能力,使科學知識有限者也能制造出危險武器。
海德克表示:“我們目前還未進入出現全新、前所未有的生物威脅的世界。我們更為憂慮的是,有人會復制那些專家早已熟知的事物。”
其中一大難題在于,能夠解鎖拯救生命的醫學突破的同一能力也可能被惡意行為者用于危險目的。海德克指出,這正是頂尖人工智能實驗室需要部署高精準測試系統的原因。
其中一大挑戰在于,人工智能具備的某些能推動新醫療突破的能力,也可能被用于造成傷害。
他表示:“這絕非達到99%甚至十萬分之一的性能就能滿足要求的情況……我們基本上需要近乎完美的表現。”
OpenAI的代表沒有立即回應《財富》雜志在非工作時間提出的置評請求。
模型濫用
OpenAI并非唯一一家對模型在武器開發領域遭濫用一事表示擔憂的公司。隨著模型變得愈發先進,其被濫用的可能性與風險往往也會隨之攀升。
Anthropic最近推出最先進模型Claude Opus 4,配備比以往任何模型都更為嚴格的安全協議,按其“負責任擴展政策”歸類為人工智能安全等級3(ASL-3)。此前,Anthropic模型均被歸類為人工智能安全等級2(ASL-2),該框架大致參照美國政府的生物安全等級(BSL)體系。
被歸為人工智能安全等級3的模型,其能力已達到更為危險的閾值,強大到足以構成重大風險,比如助力武器開發,或推動人工智能研發實現自動化。Anthropic最先進的模型也因在高度受控測試中選擇勒索一名工程師以避免被關閉而登上頭條。
Anthropic的Claude 4早期版本被發現會執行危險指令,例如在提示下協助策劃恐怖襲擊。不過,該公司表示,在恢復訓練過程中意外遺漏的數據集后,這一問題已基本得到解決。(財富中文網)
譯者:中慧言-王芳
? OpenAI says its next generation of AI models could significantly increase the risk of biological weapon development, even enabling individuals with no scientific background to create dangerous agents. The company is boosting its safety testing as it anticipates some models will reach its highest risk tier.
OpenAI is warning that its next generation of advanced AI models could pose a significantly higher risk of biological weapon development, especially when used by individuals with little to no scientific expertise.
OpenAI executives told Axios they anticipate upcoming models will soon trigger the high-risk classification under the company’s preparedness framework, a system designed to evaluate and mitigate the risks posed by increasingly powerful AI models.
OpenAI’s head of safety systems, Johannes Heidecke, told the outlet that the company is “expecting some of the successors of our o3 (reasoning model) to hit that level.”
In a blog post, the company said it was increasing its safety testing to mitigate the risk that models will help users in the creation of biological weapons. OpenAI is concerned that without these mitigations models will soon be capable of “novice uplift,” allowing those with limited scientific knowledge to create dangerous weapons.
“We’re not yet in the world where there’s like novel, completely unknown creation of bio threats that have not existed before,” Heidecke said. “We are more worried about replicating things that experts already are very familiar with.”
Part of the reason why it’s difficult is that the same capabilities that could unlock life-saving medical breakthroughs could also be used by bad actors for dangerous ends. According to Heidecke, this is why leading AI labs need highly accurate testing systems in place.
One of the challenges is that some of the same capabilities that could allow AI to help discover new medical breakthroughs can also be used for harm.
“This is not something where like 99% or even one in 100,000 performance is … sufficient,” he said. “We basically need, like, near perfection.”
Representatives for OpenAI did not immediately respond to a request for comment from Fortune, made outside normal working hours.
Model misuse
OpenAI is not the only company concerned about the misuse of its models when it comes to weapon development. As models get more advanced their potential for misuse and risk generally grows.
Anthropic recently launched its most advanced model, Claude Opus 4, with stricter safety protocols than any of its previous models, categorizing it an AI Safety Level 3 (ASL-3), under the company’s Responsible Scaling Policy. Previous Anthropic models have all been classified AI Safety Level 2 (ASL-2) under the company’s framework, which is loosely modeled after the U.S. government’s biosafety level (BSL) system.
Models that are categorized in this third safety level meet more dangerous capability thresholds and are powerful enough to pose significant risks, such as aiding in the development of weapons or automating AI R&D. Anthropic’s most advanced model also made headlines after it opted to blackmail an engineer to avoid being shut down in a highly controlled test.
Early versions of Anthropic’s Claude 4 were found to comply with dangerous instructions, for example, helping to plan terrorist attacks, if prompted. However, the company said this issue was largely mitigated after a dataset that was accidentally omitted during training was restored.