
两周前,Meta首席执行官马克·扎克伯格在其社交平台Threads上宣布启动由公司最高管理层牵头的全新“顶级计划”——Meta Compute。他借此再次强调Meta成为人工智能基础设施领域巨头的决心,并表明公司在数据中心建设竞赛中,绝无甘居人后之意。
这一新组织的设立,旨在为Meta的人工智能模型研发(终极研发目标是实现“超级智能”)提供海量算力支撑——以吉瓦为单位计量,每吉瓦算力可满足数十万户家庭的用电需求。
“Meta计划在本十年内建成数十吉瓦的算力规模,并逐步扩展至数百吉瓦乃至更高水平,”扎克伯格写道。“我们在基础设施建设过程中沉淀的工程设计经验、投资布局与合作模式,都将成为公司的战略优势。”
扎克伯格宣布,在新组织架构下,Meta资深高管桑托什·贾纳尔丹(Santosh Janardhan)将继续负责公司的技术架构、软件开发、定制芯片业务,以及庞大数据中心网络的日常建设和运营。与此同时,扎克伯格去年夏季高调招揽的人工智能专家丹尼尔·格罗斯(Daniel Gross)将领导新团队,专注于长远规划:预测数年后Meta所需的算力规模、敲定数据中心建设选址、制定稀缺芯片与能源的保障方案,以及评估这些战略投入对公司业务的潜在影响。格罗斯此前曾与OpenAI首席科学家伊尔亚·苏茨克维(Ilya Sutskever)联合创立Safe Superintelligence。
扎克伯格还宣布,任命迪娜·鲍威尔·麦考密克(Dina Powell McCormick)为Meta总裁兼副董事长,负责与各国政府建立合作关系,为全球数据中心的融资与部署提供支持。迪娜曾担任特朗普政府副国家安全战略顾问。
外界普遍认为Meta正处于追赶阶段
在部分Meta市场观察者看来,此次Meta Compute计划的发布颇为令人费解。毕竟,Meta已然是人工智能基础设施巨头。该公司位于路易斯安那州东北部的Hyperion数据中心园区动工已有一年——这座园区占地400万平方英尺(约合37万平方米),扎克伯格曾向特朗普总统坦言,其规模堪比曼哈顿下城。既然相关业务看似早已进入大规模推进阶段,Meta为何突然需要成立新顶级组织来统筹这项业务呢?
“这着实让人有些摸不着头脑,我起初也无法理解。”Moor Insights and Strategy创始人兼首席分析师帕特里克·穆尔黑德(Patrick Moorhead)坦言。他认为Meta Compute计划的发布实则面向投资者与员工,目的是向他们传递信息:在由微软、谷歌、甲骨文、OpenAI及xAI主导的人工智能巨头阵营中,Meta仍是不容小觑的强劲竞争者。“这相当于Meta在对外宣告:'这就是我们推进算力建设的战略。'”他解释道。
Bow River Capital的里克·佩德森(Rick Pederson)也认同这一观点,他认为此次公告的发布,旨在回应市场分析师的看法——目前外界普遍认为,Meta在人工智能军备竞赛中正处于追赶谷歌和OpenAI的阶段。
“通过这次公告,Meta得以阐述自身在人工智能、算力规模及基础设施建设领域的战略重心与规划。”他说道,“我猜其他大型超大规模企业可能也有类似的组织架构。只不过,扎克伯格借此机会对其进行了明确界定。”里克补充道,尽管Meta去年在人工智能基础设施领域的投入已超700亿美元,且计划在未来两年内再投入6000亿美元,但谷歌与OpenAI的投入规模至少与之持平。“因此,我认为这次公告不仅让扎克伯格有机会强调公司对算力建设的重视程度,更能详细阐释Meta的具体推进路径。”
将基础设施打造为投资组合,Meta再下重注
部分专家表示,Meta宣布推出Meta Compute计划,他们并不感到意外。“Meta正将基础设施视为投资组合进行布局,并加大投入力度,而非仅仅将其视为成本中心。”曾任职OpenAI基础设施政策主管、精品咨询公司Apeiro创始人的莱恩·迪尔格(Lane Dilg)表示。
随着人工智能热潮加速推进,Meta已不再将数据中心、图形处理器(GPU)、电力合约及定制芯片,仅仅视作支撑自身产品的“底层配套设施”,而是将其视为战略资产——其运作模式更接近资本配置者,而非传统科技公司。迪尔格指出,此举实际上意味着,Meta不仅在与其他超大规模企业竞争,更在与全球最顶尖的投资平台较量。
她解释道,在这样的行业格局下,选择丹尼尔·格罗斯共同牵头这项计划是合乎情理的。“格罗斯在构建原生人工智能与代理式人工智能平台的经验至关重要,再加上他在计算领域的专长。”她说道,并提及格罗斯与NFDG联合投资人奈特·弗里德曼(Nat Friedman,现任Meta产品负责人)共同打造超级计算机。他们的成果是名为Andromeda Cluster的算力网络,该网络由超4000块图形处理器搭建而成,他们以低于市场的价格,向旗下投资组合企业开放使用权限。
格罗斯暗示将为Meta Compute计划招揽人才。他最近在X平台上发布的一则帖子中表示,正在招募具备“深度学习、供应链管理、大宗商品交易、半导体技术、主权资产、能源领域、Excel数据分析、预测市场研究、局势监测等领域背景”的人才。这表明Meta正着手应对电力与硬件成本波动带来的风险,同时计划开展长期投资布局——其决策依据不仅涉及技术本身,还会综合考量能源市场、供应链及地缘政治因素。
资深投资人兼董事会成员乌梅什·帕德瓦尔(Umesh Padval)向《财富》杂志透露,鲍威尔·麦考密克也是Meta Compute计划的关键战略人才引进。“如今,超大规模企业正聚焦于电力获取方式,通过现金与债务融资投资电力项目,”他说道,“凭借其在金融和政界领域的背景,她将与Meta数据中心团队协同合作,推进融资并获取审批,从而加快算力基础设施的建设进程。”
批评者认为Meta的资本密集型模式将拉低回报率
然而并非所有人都支持Meta的这一举措。在Meta Compute计划发布当日,素有“大空头”之称的投资人迈克尔·伯里(Michael Burry)在社交媒体发文称:“Meta终究还是妥协了,亲手葬送了自身唯一的优势。等着看它的投资资本回报率(ROIC)暴跌吧。”伯里的警告反映出业界的担忧:Meta原本无需投入巨额资金建设实体基础设施,就能创造丰厚利润,如今却选择背离这一优势路径。他认为,Meta大举布局吉瓦级数据中心,意味着其商业模式正转向资本密集程度更高的模式,这种模式不仅会拉低公司的投资回报率,还可能让Meta逐渐蜕变为一家类似于公用事业公司的企业。
鉴于Meta已投入数百亿美元建设人工智能数据中心,且承诺追加数千亿美元用于长期基础设施建设,其发展路径显然已无逆转可能。(财富中文网)
译者:中慧言-王芳
两周前,Meta首席执行官马克·扎克伯格在其社交平台Threads上宣布启动由公司最高管理层牵头的全新“顶级计划”——Meta Compute。他借此再次强调Meta成为人工智能基础设施领域巨头的决心,并表明公司在数据中心建设竞赛中,绝无甘居人后之意。
这一新组织的设立,旨在为Meta的人工智能模型研发(终极研发目标是实现“超级智能”)提供海量算力支撑——以吉瓦为单位计量,每吉瓦算力可满足数十万户家庭的用电需求。
“Meta计划在本十年内建成数十吉瓦的算力规模,并逐步扩展至数百吉瓦乃至更高水平,”扎克伯格写道。“我们在基础设施建设过程中沉淀的工程设计经验、投资布局与合作模式,都将成为公司的战略优势。”
扎克伯格宣布,在新组织架构下,Meta资深高管桑托什·贾纳尔丹(Santosh Janardhan)将继续负责公司的技术架构、软件开发、定制芯片业务,以及庞大数据中心网络的日常建设和运营。与此同时,扎克伯格去年夏季高调招揽的人工智能专家丹尼尔·格罗斯(Daniel Gross)将领导新团队,专注于长远规划:预测数年后Meta所需的算力规模、敲定数据中心建设选址、制定稀缺芯片与能源的保障方案,以及评估这些战略投入对公司业务的潜在影响。格罗斯此前曾与OpenAI首席科学家伊尔亚·苏茨克维(Ilya Sutskever)联合创立Safe Superintelligence。
扎克伯格还宣布,任命迪娜·鲍威尔·麦考密克(Dina Powell McCormick)为Meta总裁兼副董事长,负责与各国政府建立合作关系,为全球数据中心的融资与部署提供支持。迪娜曾担任特朗普政府副国家安全战略顾问。
外界普遍认为Meta正处于追赶阶段
在部分Meta市场观察者看来,此次Meta Compute计划的发布颇为令人费解。毕竟,Meta已然是人工智能基础设施巨头。该公司位于路易斯安那州东北部的Hyperion数据中心园区动工已有一年——这座园区占地400万平方英尺(约合37万平方米),扎克伯格曾向特朗普总统坦言,其规模堪比曼哈顿下城。既然相关业务看似早已进入大规模推进阶段,Meta为何突然需要成立新顶级组织来统筹这项业务呢?
“这着实让人有些摸不着头脑,我起初也无法理解。”Moor Insights and Strategy创始人兼首席分析师帕特里克·穆尔黑德(Patrick Moorhead)坦言。他认为Meta Compute计划的发布实则面向投资者与员工,目的是向他们传递信息:在由微软、谷歌、甲骨文、OpenAI及xAI主导的人工智能巨头阵营中,Meta仍是不容小觑的强劲竞争者。“这相当于Meta在对外宣告:'这就是我们推进算力建设的战略。'”他解释道。
Bow River Capital的里克·佩德森(Rick Pederson)也认同这一观点,他认为此次公告的发布,旨在回应市场分析师的看法——目前外界普遍认为,Meta在人工智能军备竞赛中正处于追赶谷歌和OpenAI的阶段。
“通过这次公告,Meta得以阐述自身在人工智能、算力规模及基础设施建设领域的战略重心与规划。”他说道,“我猜其他大型超大规模企业可能也有类似的组织架构。只不过,扎克伯格借此机会对其进行了明确界定。”里克补充道,尽管Meta去年在人工智能基础设施领域的投入已超700亿美元,且计划在未来两年内再投入6000亿美元,但谷歌与OpenAI的投入规模至少与之持平。“因此,我认为这次公告不仅让扎克伯格有机会强调公司对算力建设的重视程度,更能详细阐释Meta的具体推进路径。”
将基础设施打造为投资组合,Meta再下重注
部分专家表示,Meta宣布推出Meta Compute计划,他们并不感到意外。“Meta正将基础设施视为投资组合进行布局,并加大投入力度,而非仅仅将其视为成本中心。”曾任职OpenAI基础设施政策主管、精品咨询公司Apeiro创始人的莱恩·迪尔格(Lane Dilg)表示。
随着人工智能热潮加速推进,Meta已不再将数据中心、图形处理器(GPU)、电力合约及定制芯片,仅仅视作支撑自身产品的“底层配套设施”,而是将其视为战略资产——其运作模式更接近资本配置者,而非传统科技公司。迪尔格指出,此举实际上意味着,Meta不仅在与其他超大规模企业竞争,更在与全球最顶尖的投资平台较量。
她解释道,在这样的行业格局下,选择丹尼尔·格罗斯共同牵头这项计划是合乎情理的。“格罗斯在构建原生人工智能与代理式人工智能平台的经验至关重要,再加上他在计算领域的专长。”她说道,并提及格罗斯与NFDG联合投资人奈特·弗里德曼(Nat Friedman,现任Meta产品负责人)共同打造超级计算机。他们的成果是名为Andromeda Cluster的算力网络,该网络由超4000块图形处理器搭建而成,他们以低于市场的价格,向旗下投资组合企业开放使用权限。
格罗斯暗示将为Meta Compute计划招揽人才。他最近在X平台上发布的一则帖子中表示,正在招募具备“深度学习、供应链管理、大宗商品交易、半导体技术、主权资产、能源领域、Excel数据分析、预测市场研究、局势监测等领域背景”的人才。这表明Meta正着手应对电力与硬件成本波动带来的风险,同时计划开展长期投资布局——其决策依据不仅涉及技术本身,还会综合考量能源市场、供应链及地缘政治因素。
资深投资人兼董事会成员乌梅什·帕德瓦尔(Umesh Padval)向《财富》杂志透露,鲍威尔·麦考密克也是Meta Compute计划的关键战略人才引进。“如今,超大规模企业正聚焦于电力获取方式,通过现金与债务融资投资电力项目,”他说道,“凭借其在金融和政界领域的背景,她将与Meta数据中心团队协同合作,推进融资并获取审批,从而加快算力基础设施的建设进程。”
批评者认为Meta的资本密集型模式将拉低回报率
然而并非所有人都支持Meta的这一举措。在Meta Compute计划发布当日,素有“大空头”之称的投资人迈克尔·伯里(Michael Burry)在社交媒体发文称:“Meta终究还是妥协了,亲手葬送了自身唯一的优势。等着看它的投资资本回报率(ROIC)暴跌吧。”伯里的警告反映出业界的担忧:Meta原本无需投入巨额资金建设实体基础设施,就能创造丰厚利润,如今却选择背离这一优势路径。他认为,Meta大举布局吉瓦级数据中心,意味着其商业模式正转向资本密集程度更高的模式,这种模式不仅会拉低公司的投资回报率,还可能让Meta逐渐蜕变为一家类似于公用事业公司的企业。
鉴于Meta已投入数百亿美元建设人工智能数据中心,且承诺追加数千亿美元用于长期基础设施建设,其发展路径显然已无逆转可能。(财富中文网)
译者:中慧言-王芳
Meta CEO Mark Zuckerberg took to his social network Threads two weeks ago to announce Meta Compute, a new “top-level initiative” led by the company’s most senior executives. In doing so, he reemphasized Meta’s commitment to being an AI infrastructure behemoth—and signaled that Meta has no intention of being an also-ran in the data center build-out race.
The new organization is designed to secure the massive amounts of computing power—measured in gigawatts, each of which could power hundreds of thousands of homes—needed for Meta’s drive to build AI models that lead to “superintelligence.”
“Meta is planning to build tens of gigawatts this decade, and hundreds of gigawatts or more over time,” Zuckerberg wrote. “How we engineer, invest, and partner to build this infrastructure will become a strategic advantage.”
Under the new structure, Zuckerberg said longtime Meta executive Santosh Janardhan will continue to run the company’s technical architecture, software, custom chips, and the day-to-day building and operation of Meta’s vast data center network. Meanwhile, Daniel Gross—one of Zuckerberg’s high-profile AI hires from last summer, who was previously cofounder of Safe Superintelligence with former OpenAI chief scientist Ilya Sutskever—will lead a new group focused on the long game: how much computing power Meta will need years from now, where it should be built, how to secure scarce chips and energy, and how to model the business impact of those bets.
Zuckerberg also announced a new Meta president and vice chair, Dina Powell McCormick, to work on developing partnerships with governments to finance and deploy data centers around the world. She was previously deputy national security advisor for strategy to President Trump.
A growing perception that Meta is playing catch-up
For some Meta-watchers, the Meta Compute announcement was puzzling. After all, Meta is already an AI infrastructure giant. It has been a year since the company broke ground on its Hyperion site—a 4-million-square-foot data center campus in northeast Louisiana that Zuckerberg famously told President Trump was roughly comparable to the size of lower Manhattan. Why, then, did Meta suddenly need to declare a new top-level organization to do what it already appears to be doing at historic scale?
“This was kind of a head scratcher, I didn’t understand it at first,” said Patrick Moorhead, founder and chief analyst at Moor Insights and Strategy. He suggested the message about Meta Compute was really meant for investors and employees, signaling that Meta remains a serious contender in a club led by Microsoft, Google, Oracle, OpenAI, and xAI. “It’s Meta saying, ‘Here is our strategy to roll this out,’” he said.
Rick Pederson of Bow River Capital agreed that the announcement responds to a growing perception among market analysts that Meta is playing catch-up with Google and OpenAI in the AI arms race.
“It was a way to discuss their focus, their intentionality surrounding the build-out of AI, computational capacity and infrastructure,” he said. “I would imagine that the other major hyperscalers have organizations something like this. But he took the opportunity to define it.” Even though the company spent over $70 billion on AI infrastructure last year and plans to spend another $600 billion in the next two years, Google and OpenAI are spending at least that amount, he added. “So I think this gave Zuckerberg a chance to talk about not just the emphasis, but also how they’re going to do it.”
Doubling down on infrastructure as an investment portfolio
Some experts said they are not at all surprised by the move to announce Meta Compute. “Meta is doubling down on infrastructure as an investment portfolio rather than simply a cost center,” said Lane Dilg, former head of infrastructure policy at OpenAI and founder of boutique advisory firm Apeiro.
As the AI boom accelerates, Meta no longer sees data centers, GPUs, power contracts, and custom chips as mere plumbing for its products. It treats them as strategic assets—more like a capital allocator than a tech company. In effect, Dilg said, Meta is positioning itself not just against other hyperscalers, but against the world’s most sophisticated investment platforms.
In that landscape, she explained, choosing Daniel Gross to co-lead the initiative makes sense. “Gross’s experience building AI-native and agentic platforms matters, paired with his compute expertise,” she said, pointing to the supercomputer he built with NFDG coinvestor Nat Friedman—now Meta’s head of products. Their effort became the Andromeda Cluster, a network of computing stockpiles totaling more than 4,000 GPUs, which they made available to their portfolio companies at below-market rates.
Gross has dropped hints about his own recruiting push for Meta Compute, saying in a recent post on X that he is hiring people with backgrounds in “deep learning, supply chains, commodities, semiconductors, sovereigns, energy, Excel, prediction markets, monitoring situations, etc.” This suggests Meta is preparing to hedge against volatility in power and hardware costs—and to make long-term bets shaped not just by technology, but by energy markets, supply chains, and geopolitics.
Powell McCormick is also a key strategic hire for the Meta Compute effort, Umesh Padval, an experienced investor and board member, told Fortune. “Hyperscalers are now focused on how to get power and are investing in power projects financed by cash and debt,” he said. “With her banking and political background, she would work with the Meta data center group to finance and get approvals to enable faster building of the compute capacity.”
Critics say Meta’s capital-intensive model would drag down returns
Not all are supportive of Meta’s move, however. In a social media post shared the day Meta Compute was announced, “Big Short” investor Michael Burry wrote: “Meta gives in, throwing away its one saving grace. Watch ROIC crash.” Burry’s warning reflects a fear that Meta is abandoning its ability to generate enormous profits without sinking vast sums into physical infrastructure. By embracing gigawatt-scale data centers, he argues, Meta is shifting toward a far more capital-intensive model—one that could drag down returns and make the company look more like a utility.
But since Meta has already poured tens of billions into AI data centers and signaled hundreds of billions more in long-term infrastructure commitments, it looks like that train left the station long ago.