
试想一下,你指示AI智能体在当天结束前将1万美元兑换成加元。智能体执行了任务,却搞砸了。它误读了参数,未经授权进行了杠杆交易,导致你的本金瞬间亏空。此时,责任该由谁承担?损失又该由谁来赔偿?
目前,没有任何一方需要为此担责。一组研究人员指出,这正是代理式人工智能时代的致命短板。
在4月8日发表的一篇论文中,来自微软研究院、哥伦比亚大学、谷歌DeepMind、Virtuals Protocol以及人工智能初创公司T54 Labs的研究人员,提出了一套名为“代理风险标准”(ARS)的全新综合性金融保护框架,旨在为AI智能体提供与传统金融交易中第三方托管、保险和清算所同等效力的保障机制。该标准是开源的,可通过T54 Labs在GitHub上获取。
T54创始人钱德勒·方(Chandler Fang)在发给《财富》杂志的电子邮件声明中表示:“我们探讨的是整个‘代理经济’,这与仅用AI智能体完成金融任务有着本质区别。”他指出,代理交易主要分为两大类:人工介入的金融交易和智能体自主交易。他表示,目前人们都聚焦“人工介入”交易,而这恰恰是问题所在,因为当前的金融生态系统除了将全部责任推给人类之外,别无他法。研究人员解释称,这一切的根源在于这项技术本身的概率性特质。
概率性问题
该团队指出,核心问题在于他们所说的“保障缺口”,其定义为“人工智能安全技术所提供的概率性可靠性,与用户委托高风险任务前所需的可执行保障之间存在脱节”。这一表述让人们联想到领导力专家杰森·怀尔德(Jason Wild)此前向《财富》杂志阐述的观点:人工智能工具的概率性特质,让各地的管理者感到困惑。T54团队写道:“由于无法限制潜在损失,用户会明智地将人工智能任务限定在低风险范围内,从而限制了基于智能体的服务的在更广泛场景中的应用。”
他们认为,模型层面的安全优化虽能降低人工智能出现故障的概率,却无法彻底消除风险。大语言模型本质上具有随机性,这意味着,无论AI智能体经过多么严苛的训练或调优,仍可能出现“幻觉”、失误。当该智能体操控着你的证券经纪账户或执行金融应用程序编程接口(API)调用时,哪怕仅出现一次失误,都可能立即造成损失。
“大多数可信的人工智能研究的核心目标都是降低故障发生的概率,”微软研究院高级研究员华文越(Wenyue Hua)表示,“这项工作至关重要,但概率永远无法等同于保障。代理风险标准采用了互补的方法:我们不追求打造完美无缺的模型,而是明确界定模型出现故障时的财务处理规则。最终我们构建了一套清算协议,为用户提供确定性的保障,而非概率性的承诺。”
研究人员提出的解决方案直接借鉴了数百年来的金融工程经验。代理风险标准引入了分层清算框架:设立托管金库,用于保管服务费,仅在任务完成并通过核验后,才向服务方划转资金;要求人工智能服务提供商在动用用户资金前,缴纳抵押担保;设置可选承保机制——由承担风险的第三方机构对人工智能故障风险进行定价,收取保费,并承诺出现问题时赔付用户。
该框架将两类人工智能任务区分开来:标准服务任务(如制作幻灯片、撰写报告)涉及的财务风险有限,因此基于第三方托管的清算机制足以覆盖风险;涉及资金交易的任务——如外汇交易、杠杆头寸、金融应用程序编程接口调用——要求智能体在结果核验前动用用户资金,在此类场景下,承保机制就变得至关重要。这与衍生品市场的运作原理完全一致:在衍生品市场中,清算所作为交易双方的中间机构,避免单次违约事件引发连锁反应。
论文通过表格形式,将代理风险标准与现有风险分摊机制进行了明确对标:建筑业采用履约保函;电子商务依托平台托管;金融市场执行保证金制度并由清算所兜底;去中心化金融(DeFi)依赖智能合约抵押。研究人员认为,AI智能体不过是下一个需要构建自身基础设施的高风险服务类别。
时机至关重要
金融监管机构已开始密切关注相关动态。美国金融业监管局(FINRA)在12月发布的《2026年监管监督报告》中,首次增设生成式人工智能章节,警告经纪商需针对人工智能“幻觉”问题制定专项处理流程,并严格监控可能“超出用户实际或预期范围及权限”运行的AI智能体。美国证券交易委员会(SEC)及其他监管机构也密切关注该领域动向。
监管机构尚未建立代理风险标准:该标准不是规则,而是协议——标准化的状态机,用于管理资金锁定、索赔申报,并在AI智能体出现故障时触发赔付。研究人员承认,代理风险标准只是庞大信任体系中的一个层级,真正的瓶颈在于为代理行为构建精准的风险定价模型。
钱德勒·方告诉《财富》杂志:“本研究是构建高层次框架的初步探索,旨在系统梳理智能体自主交易的完整流程,并对其风险评估机制进行分析。未来,我们应纳入特定细则、模型和其他研究成果,厘清如何在不同应用场景中开展风险评估。”(财富中文网)
译者:中慧言-王芳
试想一下,你指示AI智能体在当天结束前将1万美元兑换成加元。智能体执行了任务,却搞砸了。它误读了参数,未经授权进行了杠杆交易,导致你的本金瞬间亏空。此时,责任该由谁承担?损失又该由谁来赔偿?
目前,没有任何一方需要为此担责。一组研究人员指出,这正是代理式人工智能时代的致命短板。
在4月8日发表的一篇论文中,来自微软研究院、哥伦比亚大学、谷歌DeepMind、Virtuals Protocol以及人工智能初创公司T54 Labs的研究人员,提出了一套名为“代理风险标准”(ARS)的全新综合性金融保护框架,旨在为AI智能体提供与传统金融交易中第三方托管、保险和清算所同等效力的保障机制。该标准是开源的,可通过T54 Labs在GitHub上获取。
T54创始人钱德勒·方(Chandler Fang)在发给《财富》杂志的电子邮件声明中表示:“我们探讨的是整个‘代理经济’,这与仅用AI智能体完成金融任务有着本质区别。”他指出,代理交易主要分为两大类:人工介入的金融交易和智能体自主交易。他表示,目前人们都聚焦“人工介入”交易,而这恰恰是问题所在,因为当前的金融生态系统除了将全部责任推给人类之外,别无他法。研究人员解释称,这一切的根源在于这项技术本身的概率性特质。
概率性问题
该团队指出,核心问题在于他们所说的“保障缺口”,其定义为“人工智能安全技术所提供的概率性可靠性,与用户委托高风险任务前所需的可执行保障之间存在脱节”。这一表述让人们联想到领导力专家杰森·怀尔德(Jason Wild)此前向《财富》杂志阐述的观点:人工智能工具的概率性特质,让各地的管理者感到困惑。T54团队写道:“由于无法限制潜在损失,用户会明智地将人工智能任务限定在低风险范围内,从而限制了基于智能体的服务的在更广泛场景中的应用。”
他们认为,模型层面的安全优化虽能降低人工智能出现故障的概率,却无法彻底消除风险。大语言模型本质上具有随机性,这意味着,无论AI智能体经过多么严苛的训练或调优,仍可能出现“幻觉”、失误。当该智能体操控着你的证券经纪账户或执行金融应用程序编程接口(API)调用时,哪怕仅出现一次失误,都可能立即造成损失。
“大多数可信的人工智能研究的核心目标都是降低故障发生的概率,”微软研究院高级研究员华文越(Wenyue Hua)表示,“这项工作至关重要,但概率永远无法等同于保障。代理风险标准采用了互补的方法:我们不追求打造完美无缺的模型,而是明确界定模型出现故障时的财务处理规则。最终我们构建了一套清算协议,为用户提供确定性的保障,而非概率性的承诺。”
研究人员提出的解决方案直接借鉴了数百年来的金融工程经验。代理风险标准引入了分层清算框架:设立托管金库,用于保管服务费,仅在任务完成并通过核验后,才向服务方划转资金;要求人工智能服务提供商在动用用户资金前,缴纳抵押担保;设置可选承保机制——由承担风险的第三方机构对人工智能故障风险进行定价,收取保费,并承诺出现问题时赔付用户。
该框架将两类人工智能任务区分开来:标准服务任务(如制作幻灯片、撰写报告)涉及的财务风险有限,因此基于第三方托管的清算机制足以覆盖风险;涉及资金交易的任务——如外汇交易、杠杆头寸、金融应用程序编程接口调用——要求智能体在结果核验前动用用户资金,在此类场景下,承保机制就变得至关重要。这与衍生品市场的运作原理完全一致:在衍生品市场中,清算所作为交易双方的中间机构,避免单次违约事件引发连锁反应。
论文通过表格形式,将代理风险标准与现有风险分摊机制进行了明确对标:建筑业采用履约保函;电子商务依托平台托管;金融市场执行保证金制度并由清算所兜底;去中心化金融(DeFi)依赖智能合约抵押。研究人员认为,AI智能体不过是下一个需要构建自身基础设施的高风险服务类别。
时机至关重要
金融监管机构已开始密切关注相关动态。美国金融业监管局(FINRA)在12月发布的《2026年监管监督报告》中,首次增设生成式人工智能章节,警告经纪商需针对人工智能“幻觉”问题制定专项处理流程,并严格监控可能“超出用户实际或预期范围及权限”运行的AI智能体。美国证券交易委员会(SEC)及其他监管机构也密切关注该领域动向。
监管机构尚未建立代理风险标准:该标准不是规则,而是协议——标准化的状态机,用于管理资金锁定、索赔申报,并在AI智能体出现故障时触发赔付。研究人员承认,代理风险标准只是庞大信任体系中的一个层级,真正的瓶颈在于为代理行为构建精准的风险定价模型。
钱德勒·方告诉《财富》杂志:“本研究是构建高层次框架的初步探索,旨在系统梳理智能体自主交易的完整流程,并对其风险评估机制进行分析。未来,我们应纳入特定细则、模型和其他研究成果,厘清如何在不同应用场景中开展风险评估。”(财富中文网)
译者:中慧言-王芳
Imagine you tell an AI agent to convert $10,000 in U.S. dollars to Canadian dollars by end of day. The agent executes—badly. It misreads parameters, makes an unauthorized leveraged bet, and your capital evaporates. Who’s responsible? Who pays you back?
Right now, nobody has to. And that, a group of researchers argues, is the defining vulnerability of the agentic AI era.
In a paper published on April 8, researchers from Microsoft Research, Columbia University, Google DeepMind, Virtuals Protocol, and AI startup T54 Labs have proposed a sweeping new financial protection framework called the Agentic Risk Standard (ARS), designed to do for AI agents what escrow, insurance, and clearinghouses do for traditional financial transactions. The standard is open-source and available on GitHub via T54 Labs.
We are talking about an entire “agentic economy” here, T54 founder Chandler Fang told Fortune in an emailed statement; “it is very different from simply using AI agents for financial tasks.” He said there are two fundamental types of agentic transactions: human-in-the-loop financial transactions and agent-autonomous transactions. Everyone’s focus is on the human-in-the-loop stuff, he said, and that’s a real problem, because the financial ecosystem currently has no way to operate other than to defer all liability back to a human. It all comes down to the probabilistic nature of this technology, the researchers explained.
The probabilistic problem
The core problem the team identifies is what they call a “guarantee gap,” which they define as a “disconnect between the probabilistic reliability that AI safety techniques provide and the enforceable guarantees users need before delegating high-stakes tasks.” This description recalls what leadership expert Jason Wild previously told Fortune about how AI tools are probabilistic, befuddling managers everywhere. “Without a way to bound potential losses,” the T54 team wrote, “users rationally limit AI delegation to low-risk tasks, constraining the broader adoption of agent-based services.”
Model-level safety improvements, they argue, can reduce the probability of an AI failure, but cannot eliminate it. Large language models are inherently stochastic, meaning that no matter how well trained or well tuned an AI agent is, it can still hallucinate and make mistakes. When that agent is sitting on top of your brokerage account or executing financial API calls, even a single failure can produce immediate, realized loss.
“Most trustworthy AI research aims to reduce the probability of failure,” said Wenyue Hua, senior researcher at Microsoft Research. “That work is essential, but probability is not a guarantee. ARS takes a complementary approach: Instead of trying to make the model perfect, we formalize what happens financially when it isn’t. The result is a settlement protocol where user protection is deterministic, not probabilistic.”
The researchers’ solution borrows directly from centuries of financial engineering. ARS introduces a layered settlement framework: escrow vaults that hold service fees and release them only upon verified task delivery; collateral requirements that AI service providers must post before accessing user funds; and optional underwriting—a risk-bearing third party that prices the danger of an AI failure, charges a premium, and commits to reimbursing the user if things go wrong.
The framework distinguishes between two types of AI jobs: Standard service tasks—generating a slide deck, writing a report—carry limited financial exposure, so escrow-based settlement is sufficient. Tasks involving the exchange of funds—currency trading, leveraged positions, financial API calls—require the agent to access user capital before outcomes can be verified, which is where underwriting becomes essential. It is the same logic that governs derivatives markets, where clearinghouses stand between counterparties so that a single default doesn’t cascade.
The paper maps ARS explicitly against existing risk-allocation industries in a table: construction uses performance bonds; e-commerce uses platform escrow; financial markets use margin requirements and clearinghouses; and DeFi uses smart contract collateralization. AI agents, the researchers argue, are simply the next high-stakes service category that needs its own version of that infrastructure.
The timing is crucial
Financial regulators are already circling. FINRA’s 2026 regulatory oversight report, released in December, included a first-ever section on generative AI, warning broker-dealers to develop procedures specifically targeting hallucinations and to scrutinize AI agents that may act “beyond the user’s actual or intended scope and authority.” The SEC and other agencies are watching closely.
But ARS is pitched as something regulators haven’t yet built: not a set of rules, but a protocol—a standardized state machine that governs how funds are locked, how claims are filed, and how reimbursements are triggered when an AI agent fails. The researchers acknowledge ARS is one layer of a larger trust stack, and that the real bottleneck will be building accurate risk-pricing models for agentic behavior.
“This paper is the first step in setting up a high-level framework to capture the end-to-end process associated with agent-autonomous transactions and what the risk assessment looks like,” Fang told Fortune. “Further down the road, we should introduce more specific details, models, and other research to understand how we figure out risk across different use cases.”