
公众并不热衷于人工智能技术,这令OpenAI的首席执行官山姆·奥尔特曼感到沮丧。但这个问题的源头正是奥尔特曼本人以及其他人工智能公司高管:他们一边过度炒作自家技术,一边又不断提醒公众,在一个由人工智能驱动的世界里,人们的经济安全可能受到威胁。
作为人工智能技术的“首席推销员”,奥尔特曼最近在一次行业会议上表示,对人工智能的发展速度感到失望。据《纽约时报》(New York Times)报道,他抱怨道,人工智能在文化和经济中的“扩散和吸收”所遭遇的阻力,超出他的预期。《纽约时报》还援引奥尔特曼的话称:“从技术潜力来看,现在的发展速度确实慢得出乎意料。”
而在人工智能巨头之中,持类似看法的不止奥尔特曼一人。同一篇报道还援引了英伟达(Nvidia)首席执行官黄仁勋的话说,人工智能的怀疑论者“让人们不敢对人工智能进行投资”,而这些投资本来可以推动人工智能技术的进步。
这并不令人意外,因为Anthropic的首席执行官达里奥·阿莫代经常发表文章,预言未来五年人工智能将让一半白领失业。
这些高管把当前人工智能信任危机的责任归咎于别人。他们认为是公众的问题,是市场的问题,是批评者的问题。但问题的根源更为根本:人工智能领军企业违背了一条名为“邻近可能性”的市场发展原则。
该原则认为,一项创新真正流行,必须同时满足两个条件:第一,新事物本身必须稳定可靠;第二,人们必须明白自己为什么需要它。仅仅创造出一项酷炫的新技术远远不够。如果无法让公众跟上脚步,结果要么是需求疲软[比如赛格威(Segway)平衡车],要么引发反弹(例如20世纪80年代的核能争议)。
虽然人工智能的需求并不疲软,但它仍然低于支持者所期待的水平。与此同时,外界对人工智能潜在影响的抵触情绪一直在酝酿之中。
“邻近可能性”这一概念因为史蒂文·约翰逊在2010年出版的《伟大创意的诞生:创新自然史》(Where Good Ideas Come From: The Natural History of Innovation)而广为传播。历史经验表明,任何一项创新,无论是铅笔、抽水马桶、电池还是智能手机,在真正流行并改变人类工作和生活方式之前,都会经历历史积累和规律演变,在达到某个临界点后开始普及。
所谓“可能”的技术,是指那些已经存在、运行稳定,并被消费者和企业广泛采用的技术;而“尚未可能”的技术,则尚未经过验证、不够可靠、且未被目标市场真正理解。
比如,如今,面向大众市场的电动汽车已经属于“可能技术”的范畴;而每家每户的车道上停着飞行汽车,则依然属于“尚未可能”之列。
“邻近可能性”是介于这两者之间的一条狭窄地带。创新只有落在这一区间,才可能真正改变世界:既能够拓展技术边界、改变人们的习惯,又不会因为技术频繁出故障或让人感到不安而难以普及。当创新恰好进入这一“最佳位置”时,就会给用户带来惊喜体验,并催生新的消费模式,从而激发公众热情,实现快速而广泛的普及。
例如,莱特兄弟在1903年实现首次飞行时,相关的机械和理论基础,从活塞发动机到机翼空气动力学,其实早已存在。莱特兄弟所做的,只是在既有技术基础上再向前推进一步:把合适的部件组合起来,并加入一些关键性的创新见解。
而在那之前的许多年里,发明家们已经在不断尝试飞行,因此公众也逐渐相信,机器或许真的可以实现飞行的梦想。若是再往前推二十年,动力飞行对大多数人来说仍然是科幻想象。但当北卡罗来纳州的基蒂霍克(Kitty Hawk)首次成功飞行的消息传出时,全国各地的报纸争相报道,飞机很快就被兴奋的公众所接受。
回到人工智能的话题。虽然人工智能已经存在数十年,但对许多人来说,它似乎是在2022年年底OpenAI推出ChatGPT时突然闯入人们的生活。此后,人工智能的发展速度之快,几乎超过了大多数人所经历过的任何技术进步。
科技界一再向公众宣称,从我们的工作方式、职业发展,到艺术创作和政治生态,人工智能将改变一切,甚至有一天可能反过来控制人类。
这样的说法太过激进。大众市场能够理解人工智能比传统搜索更强,这属于“邻近可能性”所描述的那种跃迁,是我们从现状迈向未来可以实现的一步跨越。
但如果有人告诉我们,现在就应该拥有一整支人工智能体团队来完成一半的工作,让生产率提高十倍;或者反过来说,大家很快都会失业,这样的跨越幅度过大。而且,这种跨越还伴随着一种威胁的意味。
那么,奥尔特曼、黄仁勋以及其他人工智能行业领袖为何还会对人工智能的普及速度低于预期感到疑惑呢?
人工智能公司现在恰恰需要一剂强效“邻近可能性”的良药。技术或许正在以惊人的速度发展,但公众的接受度却没有同步。在科技产品规划中,更明智的做法往往是先找准当下的最佳位置,同时努力构建可能需要时间消化的未来。
因此,人工智能产业的领导者或许应该适当少谈“技术革命”,更专注于推出可以在当下逐步拓展边界的产品与服务,以符合人类社会能够适应的节奏。为公众描绘一条通向未来的路径,让人们愿意参与其中,而不是感到威胁,否则,来自公众乃至政策制定者的抵制可能只会越来越强烈。
本文作者凯文·马尼与麦克·丹普豪斯合著了《品类创造法则》(The Category Creation Formula)一书,同时也是品类设计顾问公司(Category Design Advisors)的联合创始人。
Fortune.com上发表的评论文章中表达的观点,仅代表作者本人的观点,不代表《财富》杂志的观点和立场。(财富中文网)
译者:刘进龙
公众并不热衷于人工智能技术,这令OpenAI的首席执行官山姆·奥尔特曼感到沮丧。但这个问题的源头正是奥尔特曼本人以及其他人工智能公司高管:他们一边过度炒作自家技术,一边又不断提醒公众,在一个由人工智能驱动的世界里,人们的经济安全可能受到威胁。
作为人工智能技术的“首席推销员”,奥尔特曼最近在一次行业会议上表示,对人工智能的发展速度感到失望。据《纽约时报》(New York Times)报道,他抱怨道,人工智能在文化和经济中的“扩散和吸收”所遭遇的阻力,超出他的预期。《纽约时报》还援引奥尔特曼的话称:“从技术潜力来看,现在的发展速度确实慢得出乎意料。”
而在人工智能巨头之中,持类似看法的不止奥尔特曼一人。同一篇报道还援引了英伟达(Nvidia)首席执行官黄仁勋的话说,人工智能的怀疑论者“让人们不敢对人工智能进行投资”,而这些投资本来可以推动人工智能技术的进步。
这并不令人意外,因为Anthropic的首席执行官达里奥·阿莫代经常发表文章,预言未来五年人工智能将让一半白领失业。
这些高管把当前人工智能信任危机的责任归咎于别人。他们认为是公众的问题,是市场的问题,是批评者的问题。但问题的根源更为根本:人工智能领军企业违背了一条名为“邻近可能性”的市场发展原则。
该原则认为,一项创新真正流行,必须同时满足两个条件:第一,新事物本身必须稳定可靠;第二,人们必须明白自己为什么需要它。仅仅创造出一项酷炫的新技术远远不够。如果无法让公众跟上脚步,结果要么是需求疲软[比如赛格威(Segway)平衡车],要么引发反弹(例如20世纪80年代的核能争议)。
虽然人工智能的需求并不疲软,但它仍然低于支持者所期待的水平。与此同时,外界对人工智能潜在影响的抵触情绪一直在酝酿之中。
“邻近可能性”这一概念因为史蒂文·约翰逊在2010年出版的《伟大创意的诞生:创新自然史》(Where Good Ideas Come From: The Natural History of Innovation)而广为传播。历史经验表明,任何一项创新,无论是铅笔、抽水马桶、电池还是智能手机,在真正流行并改变人类工作和生活方式之前,都会经历历史积累和规律演变,在达到某个临界点后开始普及。
所谓“可能”的技术,是指那些已经存在、运行稳定,并被消费者和企业广泛采用的技术;而“尚未可能”的技术,则尚未经过验证、不够可靠、且未被目标市场真正理解。
比如,如今,面向大众市场的电动汽车已经属于“可能技术”的范畴;而每家每户的车道上停着飞行汽车,则依然属于“尚未可能”之列。
“邻近可能性”是介于这两者之间的一条狭窄地带。创新只有落在这一区间,才可能真正改变世界:既能够拓展技术边界、改变人们的习惯,又不会因为技术频繁出故障或让人感到不安而难以普及。当创新恰好进入这一“最佳位置”时,就会给用户带来惊喜体验,并催生新的消费模式,从而激发公众热情,实现快速而广泛的普及。
例如,莱特兄弟在1903年实现首次飞行时,相关的机械和理论基础,从活塞发动机到机翼空气动力学,其实早已存在。莱特兄弟所做的,只是在既有技术基础上再向前推进一步:把合适的部件组合起来,并加入一些关键性的创新见解。
而在那之前的许多年里,发明家们已经在不断尝试飞行,因此公众也逐渐相信,机器或许真的可以实现飞行的梦想。若是再往前推二十年,动力飞行对大多数人来说仍然是科幻想象。但当北卡罗来纳州的基蒂霍克(Kitty Hawk)首次成功飞行的消息传出时,全国各地的报纸争相报道,飞机很快就被兴奋的公众所接受。
回到人工智能的话题。虽然人工智能已经存在数十年,但对许多人来说,它似乎是在2022年年底OpenAI推出ChatGPT时突然闯入人们的生活。此后,人工智能的发展速度之快,几乎超过了大多数人所经历过的任何技术进步。
科技界一再向公众宣称,从我们的工作方式、职业发展,到艺术创作和政治生态,人工智能将改变一切,甚至有一天可能反过来控制人类。
这样的说法太过激进。大众市场能够理解人工智能比传统搜索更强,这属于“邻近可能性”所描述的那种跃迁,是我们从现状迈向未来可以实现的一步跨越。
但如果有人告诉我们,现在就应该拥有一整支人工智能体团队来完成一半的工作,让生产率提高十倍;或者反过来说,大家很快都会失业,这样的跨越幅度过大。而且,这种跨越还伴随着一种威胁的意味。
那么,奥尔特曼、黄仁勋以及其他人工智能行业领袖为何还会对人工智能的普及速度低于预期感到疑惑呢?
人工智能公司现在恰恰需要一剂强效“邻近可能性”的良药。技术或许正在以惊人的速度发展,但公众的接受度却没有同步。在科技产品规划中,更明智的做法往往是先找准当下的最佳位置,同时努力构建可能需要时间消化的未来。
因此,人工智能产业的领导者或许应该适当少谈“技术革命”,更专注于推出可以在当下逐步拓展边界的产品与服务,以符合人类社会能够适应的节奏。为公众描绘一条通向未来的路径,让人们愿意参与其中,而不是感到威胁,否则,来自公众乃至政策制定者的抵制可能只会越来越强烈。
本文作者凯文·马尼与麦克·丹普豪斯合著了《品类创造法则》(The Category Creation Formula)一书,同时也是品类设计顾问公司(Category Design Advisors)的联合创始人。
Fortune.com上发表的评论文章中表达的观点,仅代表作者本人的观点,不代表《财富》杂志的观点和立场。(财富中文网)
译者:刘进龙
OpenAI’s Sam Altman is upset that many members of the public aren’t thrilled with AI technology. But it’s a problem that Altman and his fellow AI executives themselves created by overhyping their technology while simultaneously rattling the public about their future economic security in an AI-powered world.
As AI’s promoter-in-chief, Altman recently spoke of his disappointment with the speed of AI’s advancement at an industry conference. According to The New York Times, he complained that there was “more resistance to ‘the diffusion, the absorption’ of AI into the culture and economy than he expected.” The Times also quoted Altman as saying “Looking at what’s possible, it does feel sort of surprisingly slow.”
And he’s not alone amongst AI titans. Nvidia CEO Jensen Huang is quoted in the same story as saying that AI skeptics are “scaring people from making the investments in AI” that would make it better.
Which is no surprise, given that Anthropic CEO Dario Amodei regularly publishes essays foreseeing AI burning down half of all white collar jobs in the next five years.
The executives are blaming others for this crisis in AI confidence. It’s the public’s fault. It’s the market’s fault. It’s the critics’ fault. But the problem is more fundamental: AI’s leading companies have run afoul of a market development principle called the “adjacent possible.”
That principle asserts that innovations only truly catch on when two factors connect: One, the new thing works reliably, and two, people understand why they need it. Simply creating a cool new technology is never enough; fail to bring the public along, and you wind up with either weak demand (think Segway) or a backlash (like with nuclear power in the 1980s).
While demand for AI isn’t weak, it’s weaker than its proponents think it should be. And at the same time, a backlash against AI has been brewing over the technology’s potential impacts.
The concept of the adjacent possible was popularized in Steven Johnson’s 2010 book, Where Good Ideas Come From: The Natural History of Innovation. Historical patterns precede an explosive moment when an innovation – the pencil, the flush toilet, batteries, the smartphone – catches on and changes the way we work or live.
“Possible” technologies already exist, work well, and have been adopted by consumers and businesses. “Not-yet-possible” technologies are untested, unreliable and not yet well-understood by their target market.
Today, for instance, mass-market electric cars land in the possible. Flying cars in every driveway land in the not-yet-possible.
The adjacent possible is a thin band between those two zones. Innovations change the world when they land there, stretching boundaries and changing habits — but not so much that the technology keeps glitching or makes us feel uneasy. When an innovation hits this sweet spot, the result is user delight and new consumption patterns that create popular enthusiasm and rapid, broad adoption.
For example, when the Wright brothers first flew in 1903, all the mechanics and theories necessary — from the piston engine to wing aerodynamics — already existed. The Wrights just had to push the technology a bit further by putting the right parts together and adding some key insights of their own.
And by that point, inventors had been trying for years to fly, so the public was ready to believe a machine could deliver on the promise. Twenty years earlier, powered flight was science fiction to most people. But news of the first successful flight at Kitty Hawk was reported breathlessly by newspapers around the country, and airplanes were soon embraced by an excited public.
Which circles back to AI. While artificial intelligence has been around for decades, for much of the population it seemed to suddenly slam into their lives with the introduction of OpenAI’s ChatGPT in late 2022. AI has since advanced faster than any technology most of us have experienced.
We’re being told over and over by the tech crowd that AI is going to change everything – the way we work, our careers, our art, our politics – and might even come to control us.
It’s too much too fast. The mass market can grasp that AI is better than search. The adjacent possible would tell us that it’s a leap we can make from where we were to where we’re going.
But telling us that we should already have a team of AI agents doing half our jobs and making us ten times more productive – or alternatively that we’ll all be unemployed in the near future — is just too far of a leap. Further, it’s a leap attached to a threat.
And Altman and Huang and other AI industry leaders wonder why AI adoption is lagging their expectations?
AI companies need a strong dose of adjacent possible medicine right now. The technology may be moving at breakneck speed, but the general public is not. In tech product planning it’s always better to hit the sweet spot now while building toward a future that may take time to digest.
So AI leaders might consider dialing back the revolution and instead focus on turning out products and services today that push us into new territory at a human pace. Map out a journey into the future for us that we can buy into without feeling threatened – or risk more pushback from the public and, eventually, policymakers.
Kevin Maney is co-author, with Mike Damphousse, of The Category Creation Formula and co-founder of Category Design Advisors.
The opinions expressed in Fortune.com commentary pieces are solely the views of their authors and do not necessarily reflect the opinions and beliefs of Fortune.