
处于关系危机中的伴侣,能否从人工智能那里获得不亚于心理治疗师的支持效果?一项最新研究给出了肯定回答。
在实验中,研究人员准备了伴侣治疗情境,例如一方漠视伴侣的抑郁情绪等,然后分别收集两个支持系统的回应:一方是经验丰富的治疗师,另一方是AI聊天机器人ChatGPT。
研究共有830名参与者,男女各占一半,平均年龄为45岁,大多数都处于婚恋关系中。参与者随机接收来自治疗师或AI的回应,看看他们能否轻松判断这些回应的来源。
与研究人员(包括主要作者盖博·哈奇和来自俄亥俄州立大学(Ohio State University)等高校的心理学和精神病学项目的同事)最开始所怀疑的那样,参与者很难分辨出回应来自ChatGPT还是治疗师;他们识别治疗师回应的成功率为56.1%,而识别ChatGPT回应的成功率为51.2%。
此外,在大多数情况下,参与者更喜欢ChatGPT给出的回应。该项研究基于五个维度进行评估:对倾诉者的理解程度、共情能力的体现、治疗场景的适用性、跨文化适用性、以及是否达到优秀治疗师的应答标准。ChatGPT在对倾诉者的理解程度、共情能力展现和文化敏感性方面有显著优势。
这项研究的作者写道:“这或许预示着ChatGPT有潜力改善心理治疗过程。”论文发表于《公共科学图书馆/心理健康》(PLOS Mental Health)杂志。
具体来说,ChatGPT有助于开发不同的测试方法,开发新的心理治疗干预方法。论文的作者呼吁心理健康专家关注这一点,因为越来越多的证据表明生成式AI可能会更快被整合到治疗环境中。
多年来,专家们指出心理诊所可以利用一些AI创新提高治疗效果,包括治疗聊天机器人、自动化诊疗记录工具、自动化行政管理工具以及更智能的培训等,而且正如美国心理学会(American Psychological Association)早在2023年所写的那样,临床医生可以利用工具理解患者需求和建立信任。
来自德克萨斯州的持证心理学家兼公平技术倡议者杰西卡·杰克逊当时对美国心理学会表示:“根本问题在于,我们没有足够的治疗师。虽然心理治疗应该面向所有人,但并非所有人都需要治疗。而聊天机器人可以满足需求。”美国心理学会的论文表示,聊天机器人的培训可以填补一些心理健康问题(如睡眠问题)方面的空白,通过更经济便利的方式提供心理健康支持,而且这对于反感人类治疗师的社交焦虑患者而言是个福音。
自1966年ELIZA(首个模拟罗杰斯学派心理治疗师的聊天机器人)问世以来,研究人员关于AI能否承担治疗师角色的争论就一直持续至今。
在这方面,新研究论文的作者写道:“虽然仍有许多重要的问题亟待解决,但我们的研究结果表明答案可能是肯定的。在AI介入心理健康治疗方面,我们希望我们的工作能够激发公众和心理健康从业者思考行为准则、可行性和使用价值等重要问题,在AI列车全速前进之前筑牢安全轨道。”(财富中文网)
译者:刘进龙
审校:汪皓
处于关系危机中的伴侣,能否从人工智能那里获得不亚于心理治疗师的支持效果?一项最新研究给出了肯定回答。
在实验中,研究人员准备了伴侣治疗情境,例如一方漠视伴侣的抑郁情绪等,然后分别收集两个支持系统的回应:一方是经验丰富的治疗师,另一方是AI聊天机器人ChatGPT。
研究共有830名参与者,男女各占一半,平均年龄为45岁,大多数都处于婚恋关系中。参与者随机接收来自治疗师或AI的回应,看看他们能否轻松判断这些回应的来源。
与研究人员(包括主要作者盖博·哈奇和来自俄亥俄州立大学(Ohio State University)等高校的心理学和精神病学项目的同事)最开始所怀疑的那样,参与者很难分辨出回应来自ChatGPT还是治疗师;他们识别治疗师回应的成功率为56.1%,而识别ChatGPT回应的成功率为51.2%。
此外,在大多数情况下,参与者更喜欢ChatGPT给出的回应。该项研究基于五个维度进行评估:对倾诉者的理解程度、共情能力的体现、治疗场景的适用性、跨文化适用性、以及是否达到优秀治疗师的应答标准。ChatGPT在对倾诉者的理解程度、共情能力展现和文化敏感性方面有显著优势。
这项研究的作者写道:“这或许预示着ChatGPT有潜力改善心理治疗过程。”论文发表于《公共科学图书馆/心理健康》(PLOS Mental Health)杂志。
具体来说,ChatGPT有助于开发不同的测试方法,开发新的心理治疗干预方法。论文的作者呼吁心理健康专家关注这一点,因为越来越多的证据表明生成式AI可能会更快被整合到治疗环境中。
多年来,专家们指出心理诊所可以利用一些AI创新提高治疗效果,包括治疗聊天机器人、自动化诊疗记录工具、自动化行政管理工具以及更智能的培训等,而且正如美国心理学会(American Psychological Association)早在2023年所写的那样,临床医生可以利用工具理解患者需求和建立信任。
来自德克萨斯州的持证心理学家兼公平技术倡议者杰西卡·杰克逊当时对美国心理学会表示:“根本问题在于,我们没有足够的治疗师。虽然心理治疗应该面向所有人,但并非所有人都需要治疗。而聊天机器人可以满足需求。”美国心理学会的论文表示,聊天机器人的培训可以填补一些心理健康问题(如睡眠问题)方面的空白,通过更经济便利的方式提供心理健康支持,而且这对于反感人类治疗师的社交焦虑患者而言是个福音。
自1966年ELIZA(首个模拟罗杰斯学派心理治疗师的聊天机器人)问世以来,研究人员关于AI能否承担治疗师角色的争论就一直持续至今。
在这方面,新研究论文的作者写道:“虽然仍有许多重要的问题亟待解决,但我们的研究结果表明答案可能是肯定的。在AI介入心理健康治疗方面,我们希望我们的工作能够激发公众和心理健康从业者思考行为准则、可行性和使用价值等重要问题,在AI列车全速前进之前筑牢安全轨道。”(财富中文网)
译者:刘进龙
审校:汪皓
Can couples in distress find as much support from artificial intelligence as they would from a therapist? Yes, a new study suggests.
For the experiment, researchers prepared couples-therapy scenarios—one in which a partner was dismissing the other’s depression, for example—and then turned to two support systems for responses: a group of experienced therapists and the ChatGPT AI chatbot.
They then had 830 study participants—about half men and half women averaging 45 years old, the majority of which were in relationships—randomly receive either a therapist or AI response to see how easily they could discern the source of each.
As suspected from the outset by researchers—including lead author Gabe Hatch and colleagues from psychology and psychiatry programs at universities including the Ohio State University—participants had difficulty identifying whether responses were written by ChatGPT or by therapists; they correctly guessed that the therapist was the author 56.1% of the time and correctly guessed it was ChatGPT 51.2% of the time.
Further, participants in most cases preferred ChatGPT’s take on the matter at hand. That was based on five factors: whether the response understood the speaker, showed empathy, was appropriate for the therapy setting, was relevant for various cultural backgrounds, and was something a good therapist would say. ChatGPT came out ahead of human therapists particularly around understanding the speaker, showing empathy, and showing cultural competence.
“This may be an early indication that ChatGPT has the potential to improve psychotherapeutic processes,” the authors of the study, published in PLOS Mental Health, wrote.
Specifically, it could lead to the development of different ways to test and create psychotherapeutic interventions—something the authors call for mental health experts to pay attention to, given the mounting evidence that generative AI could be integrated into therapeutic settings sooner rather than later.
For years, experts have noted that psychology practice could do well with some AI innovations—including therapeutic chatbots, tools that automate note taking and other administrative tasks, and more intelligent trainings—as well as clinicians had tools they could understand and trust, the American Psychological Association wrote back in 2023.
“The bottom line is we don’t have enough providers,” Jessica Jackson, a licensed psychologist and equitable technology advocate based in Texas, told APA for that story. “While therapy should be for everyone, not everyone needs it. The chatbots can fill a need.” Training from a chatbot, the piece noted, could fill in gaps for some mental health concerns (such as sleep problems), make mental health support more affordable and accessible, and be a great tool for those with social anxiety who may find human therapists off-putting.
And since the 1966 invention of ELIZA—a chatbot programmed to respond as a Rogerian psychotherapist—researchers have debated whether AI could play the role of a therapist.
Regarding that, the new study authors write, “Although there are still many important lingering questions, our findings indicate the answer may be ‘Yes.’ We hope our work galvanizes both the public and Mental Practitioners to ask important questions about the ethics, feasibility, and utility of integrating AI and mental health treatment, before the AI train leaves the station.”