
過去一周,AI領域依舊動蕩不斷:Anthropic因擔憂網絡安全風險,決定暫不發布全新的Claude Mythos模型(同時成立聯盟,利用該模型的預覽版本來強化網絡安全防御);Meta在聘請汪滔加盟后發布首個AI模型;此外,市場對OpenAI即將推出的新模型“Spud”的預期持續升溫。
當前,大多數AI模型都依賴英偉達GPU運行。這些AI芯片技術復雜、價格高昂(單價超過3萬美元),是模型訓練和推理的核心引擎。但放眼整個行業,獲取GPU芯片仍然是繞不開的瓶頸。例如,OpenAI總裁格雷格·布羅克曼就曾形容,公司內部的GPU分配過程堪稱“痛苦不堪”。
而在本周于舊金山舉行的HumanX大會上,筆者發現,即便在英偉達(Nvidia)內部,GPU同樣是稀缺資源。
筆者采訪了英偉達應用深度學習研究負責人布萊恩·卡坦扎羅。他帶領的團隊從事AI驅動的圖形、語音識別和仿真等方向的研究。早在2010年代早中期,卡坦扎羅就是最早注意到研究人員開始搶購英偉達GPU用于訓練AI模型的人之一,這一趨勢也促使公司首席執行官黃仁勛加碼AI布局,為英偉達如今的史詩級增長奠定了基礎。
但如今,即便是卡坦扎羅的團隊也難以獲得充足的GPU。他表示:“我們的團隊在工作中高度依賴AI,他們最大的抱怨就是額度不夠。他們想要更多GPU。”
“效率本身也是一種智能”
事實上,他表示,自己目前的一項主要工作就是盡可能為團隊爭取更多算力資源。他說道:“我們都面臨供應限制。黃仁勛會說,‘抱歉,布萊恩,那些芯片都賣光了。’我們只能在這樣的限制下開展工作。”
卡坦扎羅負責的項目之一,是帶領團隊開發英偉達的Nemotron系列模型。這是一組開源模型,用戶可以自由下載、使用、研究或修改。需要說明的是,英偉達并不打算在模型領域與OpenAI、Anthropic等公司正面競爭。相反,其打造這些模型旨在強化開發者生態系統,使其依賴英偉達的軟硬件體系。
Nemotron系列模型以極高的GPU利用效率著稱。卡坦扎羅表示,正是英偉達內部也面臨GPU短缺,才倒逼團隊不斷提升Nemotron模型的效率。他表示:“在供應受限的環境中,效率本身也是一種智能。”
不再只是“科研項目”
不過,令人意外的是,提升效率并未損害商業利益。卡坦扎羅表示,這是“杰文斯悖論”在起作用:當某件事變得更高效時,需求反而往往會激增。他表示:“當某件事物的效率提升時,人們總會找到各種新的使用方式。”
不過他也承認,Nemotron在英偉達內部關注度的提升,同樣幫助團隊獲得了更多資源。“我們做這個項目已經很久了,但直到最近六個月才真正受到重視。隨著公司內部越來越理解這項工作的重要性,溝通變得更順暢、協作更緊密,公司也給予我們更多支持。”
他補充說,英偉達已經意識到,不能再對AI生態系統采取“放手不管”的態度。過去,公司可以依賴其他企業開發模型和應用來帶動芯片需求。但如今,隨著AI競爭加劇、芯片供應緊張,英偉達認為自己應當在生態系統發展中扮演更積極的角色。
他表示:“過去,有人認為我們可以讓生態系統自行發展。但現在很明顯,英偉達需要承擔更重要的角色——Nemotron帶來的既是責任,也是機遇。”
這種定位也有助于提升Nemotron在英偉達內部的地位,畢竟各團隊都在爭奪稀缺的GPU資源。卡坦扎羅表示:“這已經不是一個科研項目。這不僅僅是我為團隊爭取資源的問題,還關系到英偉達的未來。”(財富中文網)
譯者:劉進龍
審校:汪皓
過去一周,AI領域依舊動蕩不斷:Anthropic因擔憂網絡安全風險,決定暫不發布全新的Claude Mythos模型(同時成立聯盟,利用該模型的預覽版本來強化網絡安全防御);Meta在聘請汪滔加盟后發布首個AI模型;此外,市場對OpenAI即將推出的新模型“Spud”的預期持續升溫。
當前,大多數AI模型都依賴英偉達GPU運行。這些AI芯片技術復雜、價格高昂(單價超過3萬美元),是模型訓練和推理的核心引擎。但放眼整個行業,獲取GPU芯片仍然是繞不開的瓶頸。例如,OpenAI總裁格雷格·布羅克曼就曾形容,公司內部的GPU分配過程堪稱“痛苦不堪”。
而在本周于舊金山舉行的HumanX大會上,筆者發現,即便在英偉達(Nvidia)內部,GPU同樣是稀缺資源。
筆者采訪了英偉達應用深度學習研究負責人布萊恩·卡坦扎羅。他帶領的團隊從事AI驅動的圖形、語音識別和仿真等方向的研究。早在2010年代早中期,卡坦扎羅就是最早注意到研究人員開始搶購英偉達GPU用于訓練AI模型的人之一,這一趨勢也促使公司首席執行官黃仁勛加碼AI布局,為英偉達如今的史詩級增長奠定了基礎。
但如今,即便是卡坦扎羅的團隊也難以獲得充足的GPU。他表示:“我們的團隊在工作中高度依賴AI,他們最大的抱怨就是額度不夠。他們想要更多GPU。”
“效率本身也是一種智能”
事實上,他表示,自己目前的一項主要工作就是盡可能為團隊爭取更多算力資源。他說道:“我們都面臨供應限制。黃仁勛會說,‘抱歉,布萊恩,那些芯片都賣光了。’我們只能在這樣的限制下開展工作。”
卡坦扎羅負責的項目之一,是帶領團隊開發英偉達的Nemotron系列模型。這是一組開源模型,用戶可以自由下載、使用、研究或修改。需要說明的是,英偉達并不打算在模型領域與OpenAI、Anthropic等公司正面競爭。相反,其打造這些模型旨在強化開發者生態系統,使其依賴英偉達的軟硬件體系。
Nemotron系列模型以極高的GPU利用效率著稱。卡坦扎羅表示,正是英偉達內部也面臨GPU短缺,才倒逼團隊不斷提升Nemotron模型的效率。他表示:“在供應受限的環境中,效率本身也是一種智能。”
不再只是“科研項目”
不過,令人意外的是,提升效率并未損害商業利益。卡坦扎羅表示,這是“杰文斯悖論”在起作用:當某件事變得更高效時,需求反而往往會激增。他表示:“當某件事物的效率提升時,人們總會找到各種新的使用方式。”
不過他也承認,Nemotron在英偉達內部關注度的提升,同樣幫助團隊獲得了更多資源。“我們做這個項目已經很久了,但直到最近六個月才真正受到重視。隨著公司內部越來越理解這項工作的重要性,溝通變得更順暢、協作更緊密,公司也給予我們更多支持。”
他補充說,英偉達已經意識到,不能再對AI生態系統采取“放手不管”的態度。過去,公司可以依賴其他企業開發模型和應用來帶動芯片需求。但如今,隨著AI競爭加劇、芯片供應緊張,英偉達認為自己應當在生態系統發展中扮演更積極的角色。
他表示:“過去,有人認為我們可以讓生態系統自行發展。但現在很明顯,英偉達需要承擔更重要的角色——Nemotron帶來的既是責任,也是機遇。”
這種定位也有助于提升Nemotron在英偉達內部的地位,畢竟各團隊都在爭奪稀缺的GPU資源。卡坦扎羅表示:“這已經不是一個科研項目。這不僅僅是我為團隊爭取資源的問題,還關系到英偉達的未來。”(財富中文網)
譯者:劉進龍
審校:汪皓
It’s been another one of those wild weeks in AI, with Anthropic electing not to release its new Claude Mythos model because of concerns about the cybersecurity risks it poses (and forming a coalition to use a preview version of the model to bolster cybersecurity defenses); Meta releasing its first AI model since hiring Alexandr Wang; and mounting expectations about OpenAI’s upcoming new “Spud” model.
Most of these AI models run on Nvidia GPUs, the sophisticated and expensive AI chips (at over $30,000 a pop) that power their training and output. But across the industry, access to those chips remains a bottleneck. OpenAI president Greg Brockman, for example, has said allocating GPUs at OpenAI is “pain and suffering.”
This week, at the HumanX conference in San Francisco, I discovered that even inside Nvidia, GPUs are scarce.
I sat down with Bryan Catanzaro, who leads applied deep learning research at Nvidia, overseeing teams working on AI-driven graphics, speech recognition, and simulation. Catanzaro was also among the first, back in the early-to-mid 2010s, to notice researchers snapping up Nvidia GPUs to train AI models—a signal that helped push CEO Jensen Huang to double down on AI, setting the stage for the company’s now-historic run.
Today, though, even Catanzaro’s teams are struggling to access enough GPUs. “My team uses AI very deeply in our work, and their primary complaint is they want higher limits,” Catanzaro told me. “They want more GPUs.”
“Efficiency is also intelligence”
In fact, he said one of his main jobs now is simply trying to secure more compute for his teams. “We’re all supply constrained,” he said. “Jensen will say, ‘I’m sorry, Bryan, but those are sold.’ We operate within those constraints.”
One of Catanzaro’s projects has been leading the team building Nvidia’s Nemotron, a family of models that are open source—meaning users can freely download them to use, study, or modify. To be clear, Nvidia isn’t trying to compete in the model-building race with the likes of OpenAI and Anthropic. Instead, it’s building them to strengthen a developer ecosystem that remains tied to Nvidia hardware and software.
The Nemotron models are known for being particularly GPU-efficient. And Catanzaro said it’s the very constraints on GPU access at Nvidia itself that is driving the push to make Nemotron models more efficient. “In a supply-constrained world, efficiency is also intelligence,” he said.
No longer a science project
But surprisingly, efficiency isn’t bad for business. Catanzaro said it was Jevons Paradox at work: When something becomes more efficient, demand often surges. “People find all sorts of new ways to use a thing when it gets more efficient,” he said.
Still, he acknowledged that Nemotron’s growing visibility inside Nvidia has also helped unlock more resources. “We’ve been working on [Nemotron] for a long time, but it’s really only in the past six months that it’s gotten more attention. As people inside Nvidia better understand the importance of this work, you get better storytelling, better collaboration, and more support across the company.”
Nvidia has realized, he added, that it can no longer take a hands-off approach to the AI ecosystem. In the past, Nvidia could rely on others to build the models and applications that drove demand for its chips. Now, as AI becomes more competitive and chip-constrained, the company sees a more active role for itself in shaping how that ecosystem develops.
“In the past, some people felt like we could just let the ecosystem take care of itself,” he said. “Now it’s much more obvious that Nvidia has a bigger role to play—a real responsibility and opportunity with Nemotron.”
That framing also helps elevate the Nemotron work inside Nvidia, where teams are competing for scarce GPU resources. “This isn’t a science project,” Catanzaro said. “It’s not just me asking for resources for my team. This is about Nvidia’s future.”
AI IN THE NEWS
The pro-Iran meme machine trolling Trump with AI Lego cartoons. A new report from Wired describes how a group of young pro-Iranian creators called Explosive Media is using AI-generated, Lego-style videos to spread sophisticated, viral propaganda during the current conflict, reaching millions across TikTok, X, and Instagram. Unlike traditional state messaging, the videos blend humor, internet-savvy cultural references, and simplified storytelling to resonate with American audiences, even incorporating memes and English-language rap. Researchers say the strategy is effective because it distills complex geopolitical events into highly shareable content while tapping into existing disaffection in the U.S., illustrating how AI tools are enabling a new kind of “slopaganda” war—where influence campaigns are faster, more targeted, and far more culturally fluent than in the past.
Amazon's Andy Jassy defends Amazon’s $200B spending spree. GeekWire reported on Amazon CEO Andy Jassy's latest shareholder letter, which revealed that AWS’s AI business has already reached a $15 billion annual revenue run rate, which Jassy argued means demand is strong enough to justify roughly $200 billion in planned capex. Jassy framed AI as a “once-in-a-lifetime” opportunity and positioned Amazon squarely in the middle of the current AI “land rush,” pointing to surging demand for its custom chips like Trainium—some of which are already largely sold out years in advance—as well as interest from customers eager to secure future capacity. The letter makes clear that Amazon is betting aggressively on owning more of the AI stack, from infrastructure to chips to potentially selling those capabilities externally.
OpenAI pauses Stargate UK data center, citing energy costs. According to Bloomberg, OpenAI is pausing its planned Stargate data center project in the UK, highlighting how even the most aggressive AI infrastructure buildouts are running up against real-world constraints like energy costs and regulation. The move comes as the company reins in spending ahead of a potential IPO and narrows focus to its core ChatGPT business amid intensifying competition from Anthropic and Google. While OpenAI says it still sees long-term potential in the UK, the decision underscores a broader reality: Massive AI infrastructure bets—from Texas to Norway to the UAE—are increasingly shaped not just by ambition, but by economics, geopolitics, and access to affordable power.
EYE ON AI NUMBERS
That's how many executives say their AI strategy is more about optics than any actual internal guidance, according to Writer's new 2026 Enterprise AI Adoption Report, which surveyed 2,400 knowledge workers including 1,200 C-suite executives and 1,200 employees. In addition, 39% have no plan for how AI actually drives revenue. Yet, 69% are planning layoffs this year.
In a LinkedIn post, Writer CEO May Habib called this trend "‘AI theater’ at its worst," adding "this strategy vacuum up top is literally tearing companies apart."