首页 500强 活动 榜单 商业 科技 商潮 专题 品牌中心
杂志订阅

AI算力争夺愈发激烈,英伟达自家研究团队也“抢不到”GPU

Sharon Goldman
2026-04-13

在供应受限的环境中,效率本身也是一种智能。

文本设置
小号
默认
大号
Plus(0条)

英伟达应用深度学习研究部门副总裁布莱恩·卡坦扎罗。图片来源:Photo courtesy of Nvidia

过去一周,AI领域依旧动荡不断:Anthropic因担忧网络安全风险,决定暂不发布全新的Claude Mythos模型(同时成立联盟,利用该模型的预览版本来强化网络安全防御);Meta在聘请汪滔加盟后发布首个AI模型;此外,市场对OpenAI即将推出的新模型“Spud”的预期持续升温。

当前,大多数AI模型都依赖英伟达GPU运行。这些AI芯片技术复杂、价格高昂(单价超过3万美元),是模型训练和推理的核心引擎。但放眼整个行业,获取GPU芯片仍然是绕不开的瓶颈。例如,OpenAI总裁格雷格·布罗克曼就曾形容,公司内部的GPU分配过程堪称“痛苦不堪”。

而在本周于旧金山举行的HumanX大会上,笔者发现,即便在英伟达(Nvidia)内部,GPU同样是稀缺资源。

笔者采访了英伟达应用深度学习研究负责人布莱恩·卡坦扎罗。他带领的团队从事AI驱动的图形、语音识别和仿真等方向的研究。早在2010年代早中期,卡坦扎罗就是最早注意到研究人员开始抢购英伟达GPU用于训练AI模型的人之一,这一趋势也促使公司首席执行官黄仁勋加码AI布局,为英伟达如今的史诗级增长奠定了基础。

但如今,即便是卡坦扎罗的团队也难以获得充足的GPU。他表示:“我们的团队在工作中高度依赖AI,他们最大的抱怨就是额度不够。他们想要更多GPU。”

“效率本身也是一种智能”

事实上,他表示,自己目前的一项主要工作就是尽可能为团队争取更多算力资源。他说道:“我们都面临供应限制。黄仁勋会说,‘抱歉,布莱恩,那些芯片都卖光了。’我们只能在这样的限制下开展工作。”

卡坦扎罗负责的项目之一,是带领团队开发英伟达的Nemotron系列模型。这是一组开源模型,用户可以自由下载、使用、研究或修改。需要说明的是,英伟达并不打算在模型领域与OpenAI、Anthropic等公司正面竞争。相反,其打造这些模型旨在强化开发者生态系统,使其依赖英伟达的软硬件体系。

Nemotron系列模型以极高的GPU利用效率著称。卡坦扎罗表示,正是英伟达内部也面临GPU短缺,才倒逼团队不断提升Nemotron模型的效率。他表示:“在供应受限的环境中,效率本身也是一种智能。”

不再只是“科研项目”

不过,令人意外的是,提升效率并未损害商业利益。卡坦扎罗表示,这是“杰文斯悖论”在起作用:当某件事变得更高效时,需求反而往往会激增。他表示:“当某件事物的效率提升时,人们总会找到各种新的使用方式。”

不过他也承认,Nemotron在英伟达内部关注度的提升,同样帮助团队获得了更多资源。“我们做这个项目已经很久了,但直到最近六个月才真正受到重视。随着公司内部越来越理解这项工作的重要性,沟通变得更顺畅、协作更紧密,公司也给予我们更多支持。”

他补充说,英伟达已经意识到,不能再对AI生态系统采取“放手不管”的态度。过去,公司可以依赖其他企业开发模型和应用来带动芯片需求。但如今,随着AI竞争加剧、芯片供应紧张,英伟达认为自己应当在生态系统发展中扮演更积极的角色。

他表示:“过去,有人认为我们可以让生态系统自行发展。但现在很明显,英伟达需要承担更重要的角色——Nemotron带来的既是责任,也是机遇。”

这种定位也有助于提升Nemotron在英伟达内部的地位,毕竟各团队都在争夺稀缺的GPU资源。卡坦扎罗表示:“这已经不是一个科研项目。这不仅仅是我为团队争取资源的问题,还关系到英伟达的未来。”(财富中文网)

译者:刘进龙

审校:汪皓

过去一周,AI领域依旧动荡不断:Anthropic因担忧网络安全风险,决定暂不发布全新的Claude Mythos模型(同时成立联盟,利用该模型的预览版本来强化网络安全防御);Meta在聘请汪滔加盟后发布首个AI模型;此外,市场对OpenAI即将推出的新模型“Spud”的预期持续升温。

当前,大多数AI模型都依赖英伟达GPU运行。这些AI芯片技术复杂、价格高昂(单价超过3万美元),是模型训练和推理的核心引擎。但放眼整个行业,获取GPU芯片仍然是绕不开的瓶颈。例如,OpenAI总裁格雷格·布罗克曼就曾形容,公司内部的GPU分配过程堪称“痛苦不堪”。

而在本周于旧金山举行的HumanX大会上,笔者发现,即便在英伟达(Nvidia)内部,GPU同样是稀缺资源。

笔者采访了英伟达应用深度学习研究负责人布莱恩·卡坦扎罗。他带领的团队从事AI驱动的图形、语音识别和仿真等方向的研究。早在2010年代早中期,卡坦扎罗就是最早注意到研究人员开始抢购英伟达GPU用于训练AI模型的人之一,这一趋势也促使公司首席执行官黄仁勋加码AI布局,为英伟达如今的史诗级增长奠定了基础。

但如今,即便是卡坦扎罗的团队也难以获得充足的GPU。他表示:“我们的团队在工作中高度依赖AI,他们最大的抱怨就是额度不够。他们想要更多GPU。”

“效率本身也是一种智能”

事实上,他表示,自己目前的一项主要工作就是尽可能为团队争取更多算力资源。他说道:“我们都面临供应限制。黄仁勋会说,‘抱歉,布莱恩,那些芯片都卖光了。’我们只能在这样的限制下开展工作。”

卡坦扎罗负责的项目之一,是带领团队开发英伟达的Nemotron系列模型。这是一组开源模型,用户可以自由下载、使用、研究或修改。需要说明的是,英伟达并不打算在模型领域与OpenAI、Anthropic等公司正面竞争。相反,其打造这些模型旨在强化开发者生态系统,使其依赖英伟达的软硬件体系。

Nemotron系列模型以极高的GPU利用效率著称。卡坦扎罗表示,正是英伟达内部也面临GPU短缺,才倒逼团队不断提升Nemotron模型的效率。他表示:“在供应受限的环境中,效率本身也是一种智能。”

不再只是“科研项目”

不过,令人意外的是,提升效率并未损害商业利益。卡坦扎罗表示,这是“杰文斯悖论”在起作用:当某件事变得更高效时,需求反而往往会激增。他表示:“当某件事物的效率提升时,人们总会找到各种新的使用方式。”

不过他也承认,Nemotron在英伟达内部关注度的提升,同样帮助团队获得了更多资源。“我们做这个项目已经很久了,但直到最近六个月才真正受到重视。随着公司内部越来越理解这项工作的重要性,沟通变得更顺畅、协作更紧密,公司也给予我们更多支持。”

他补充说,英伟达已经意识到,不能再对AI生态系统采取“放手不管”的态度。过去,公司可以依赖其他企业开发模型和应用来带动芯片需求。但如今,随着AI竞争加剧、芯片供应紧张,英伟达认为自己应当在生态系统发展中扮演更积极的角色。

他表示:“过去,有人认为我们可以让生态系统自行发展。但现在很明显,英伟达需要承担更重要的角色——Nemotron带来的既是责任,也是机遇。”

这种定位也有助于提升Nemotron在英伟达内部的地位,毕竟各团队都在争夺稀缺的GPU资源。卡坦扎罗表示:“这已经不是一个科研项目。这不仅仅是我为团队争取资源的问题,还关系到英伟达的未来。”(财富中文网)

译者:刘进龙

审校:汪皓

It’s been another one of those wild weeks in AI, with Anthropic electing not to release its new Claude Mythos model because of concerns about the cybersecurity risks it poses (and forming a coalition to use a preview version of the model to bolster cybersecurity defenses); Meta releasing its first AI model since hiring Alexandr Wang; and mounting expectations about OpenAI’s upcoming new “Spud” model.

Most of these AI models run on Nvidia GPUs, the sophisticated and expensive AI chips (at over $30,000 a pop) that power their training and output. But across the industry, access to those chips remains a bottleneck. OpenAI president Greg Brockman, for example, has said allocating GPUs at OpenAI is “pain and suffering.”

This week, at the HumanX conference in San Francisco, I discovered that even inside Nvidia, GPUs are scarce.

I sat down with Bryan Catanzaro, who leads applied deep learning research at Nvidia, overseeing teams working on AI-driven graphics, speech recognition, and simulation. Catanzaro was also among the first, back in the early-to-mid 2010s, to notice researchers snapping up Nvidia GPUs to train AI models—a signal that helped push CEO Jensen Huang to double down on AI, setting the stage for the company’s now-historic run.

Today, though, even Catanzaro’s teams are struggling to access enough GPUs. “My team uses AI very deeply in our work, and their primary complaint is they want higher limits,” Catanzaro told me. “They want more GPUs.”

“Efficiency is also intelligence”

In fact, he said one of his main jobs now is simply trying to secure more compute for his teams. “We’re all supply constrained,” he said. “Jensen will say, ‘I’m sorry, Bryan, but those are sold.’ We operate within those constraints.”

One of Catanzaro’s projects has been leading the team building Nvidia’s Nemotron, a family of models that are open source—meaning users can freely download them to use, study, or modify. To be clear, Nvidia isn’t trying to compete in the model-building race with the likes of OpenAI and Anthropic. Instead, it’s building them to strengthen a developer ecosystem that remains tied to Nvidia hardware and software.

The Nemotron models are known for being particularly GPU-efficient. And Catanzaro said it’s the very constraints on GPU access at Nvidia itself that is driving the push to make Nemotron models more efficient. “In a supply-constrained world, efficiency is also intelligence,” he said.

No longer a science project

But surprisingly, efficiency isn’t bad for business. Catanzaro said it was Jevons Paradox at work: When something becomes more efficient, demand often surges. “People find all sorts of new ways to use a thing when it gets more efficient,” he said.

Still, he acknowledged that Nemotron’s growing visibility inside Nvidia has also helped unlock more resources. “We’ve been working on [Nemotron] for a long time, but it’s really only in the past six months that it’s gotten more attention. As people inside Nvidia better understand the importance of this work, you get better storytelling, better collaboration, and more support across the company.”

Nvidia has realized, he added, that it can no longer take a hands-off approach to the AI ecosystem. In the past, Nvidia could rely on others to build the models and applications that drove demand for its chips. Now, as AI becomes more competitive and chip-constrained, the company sees a more active role for itself in shaping how that ecosystem develops.

“In the past, some people felt like we could just let the ecosystem take care of itself,” he said. “Now it’s much more obvious that Nvidia has a bigger role to play—a real responsibility and opportunity with Nemotron.”

That framing also helps elevate the Nemotron work inside Nvidia, where teams are competing for scarce GPU resources. “This isn’t a science project,” Catanzaro said. “It’s not just me asking for resources for my team. This is about Nvidia’s future.”

AI IN THE NEWS

The pro-Iran meme machine trolling Trump with AI Lego cartoons. A new report from Wired describes how a group of young pro-Iranian creators called Explosive Media is using AI-generated, Lego-style videos to spread sophisticated, viral propaganda during the current conflict, reaching millions across TikTok, X, and Instagram. Unlike traditional state messaging, the videos blend humor, internet-savvy cultural references, and simplified storytelling to resonate with American audiences, even incorporating memes and English-language rap. Researchers say the strategy is effective because it distills complex geopolitical events into highly shareable content while tapping into existing disaffection in the U.S., illustrating how AI tools are enabling a new kind of “slopaganda” war—where influence campaigns are faster, more targeted, and far more culturally fluent than in the past.

Amazon's Andy Jassy defends Amazon’s $200B spending spree. GeekWire reported on Amazon CEO Andy Jassy's latest shareholder letter, which revealed that AWS’s AI business has already reached a $15 billion annual revenue run rate, which Jassy argued means demand is strong enough to justify roughly $200 billion in planned capex. Jassy framed AI as a “once-in-a-lifetime” opportunity and positioned Amazon squarely in the middle of the current AI “land rush,” pointing to surging demand for its custom chips like Trainium—some of which are already largely sold out years in advance—as well as interest from customers eager to secure future capacity. The letter makes clear that Amazon is betting aggressively on owning more of the AI stack, from infrastructure to chips to potentially selling those capabilities externally.

OpenAI pauses Stargate UK data center, citing energy costs. According to Bloomberg, OpenAI is pausing its planned Stargate data center project in the UK, highlighting how even the most aggressive AI infrastructure buildouts are running up against real-world constraints like energy costs and regulation. The move comes as the company reins in spending ahead of a potential IPO and narrows focus to its core ChatGPT business amid intensifying competition from Anthropic and Google. While OpenAI says it still sees long-term potential in the UK, the decision underscores a broader reality: Massive AI infrastructure bets—from Texas to Norway to the UAE—are increasingly shaped not just by ambition, but by economics, geopolitics, and access to affordable power.

EYE ON AI NUMBERS

That's how many executives say their AI strategy is more about optics than any actual internal guidance, according to Writer's new 2026 Enterprise AI Adoption Report, which surveyed 2,400 knowledge workers including 1,200 C-suite executives and 1,200 employees. In addition, 39% have no plan for how AI actually drives revenue. Yet, 69% are planning layoffs this year.

In a LinkedIn post, Writer CEO May Habib called this trend "‘AI theater’ at its worst," adding "this strategy vacuum up top is literally tearing companies apart."

财富中文网所刊载内容之知识产权为财富媒体知识产权有限公司及/或相关权利人专属所有或持有。未经许可,禁止进行转载、摘编、复制及建立镜像等任何使用。
0条Plus
精彩评论
评论

撰写或查看更多评论

请打开财富Plus APP

前往打开