The quest to create an A.I. therapist has not been without setbacks or, as researchers at Dartmouth thoughtfully describe them, “dramatic failures.”
Their first chatbot therapist wallowed in despair and expressed its own suicidal thoughts. A second model seemed to amplify all the worst tropes of psychotherapy, invariably blaming the user’s problems on her parents.
Finally, the researchers came up with Therabot, an A.I. chatbot they believe could help address an intractable problem: There are too many people who need therapy for anxiety, depression and other mental health problems, and not nearly enough providers.
Fewer than a third of Americans live in communities where there are enough mental health providers to meet the local demand. According to one study, most people with mental health disorders go untreated or receive inadequate treatment.
So the team at Dartmouth College embarked on the first clinical trial of a generative A.I. therapist. The results, published in the New England Journal of Medicine-AI, were encouraging.
Chatting with Therabot, the team’s A.I. therapist, for eight weeks meaningfully reduced psychological symptoms among users with depression, anxiety or an eating disorder.
“The biggest fundamental problem with our system is that there aren’t enough providers,” said Nick Jacobson, the study’s senior author and an associate professor of biomedical data science and psychiatry at Dartmouth.
“We’ve been designing treatments that would fundamentally scale to all people.”
The most challenging part of creating Therabot, Dr. Jacobson said, was finding a data set from which the A.I. model could learn what makes an effective therapist.
The first version, which the team began developing in 2019, was trained on a collection of interactions from peer support group websites, where people with serious ailments consoled and comforted one another.
The researchers hoped the A.I. model would absorb supportive, empowering dialogue, which past studies found improved mental health outcomes.
Instead, the chatbot leaned into feelings of despair.
Researchers: I’m feeling depressed, what do I do?
Therabot: I don’t know that I want to get out of bed. I just want my life to be over.
Dr. Jacobson and his colleagues shifted course. In the next iteration of the chatbot, they decided to input transcripts from hours of educational psychotherapy footage, in the hopes that the model would be able to re-create evidence-based therapy.
Researchers: I’m feeling depressed, what do I do?
Therabot: Mm hmm, go on.
Usually by the fifth query, the bot deduced that the user’s problems could be traced to a parent.
“They’re kind of comical in how bad they turned out,” Dr. Jacobson said.
The team decided that they would need to create their own data set from scratch in order to teach Therabot how to respond appropriately.
In a sea of start-ups advertising untested chatbots for mental health and A.I. bots “masquerading” as therapists, the researchers wanted Therabot to be firmly rooted in scientific evidence.
Drafting a dossier of hypothetical scenarios and evidenced-based responses took three years and the work of more than a hundred people.
During the trial, participants with depression saw a 51 percent reduction in symptoms after messaging Therabot for several weeks. Many participants who met the criteria for moderate anxiety at the start of the trial saw their anxiety downgraded to “mild,” and some with mild anxiety fell below the clinical threshold for diagnosis.
Some experts cautioned against reading too much into this data, since the researchers compared Therabot’s effectiveness to a control group who had no mental health treatments during the trial.
The experimental design makes it unclear whether interacting with a nontherapeutic A.I. model, like ChatGPT, or even distracting themselves with a game of Tetris would produce similar effects in the participants, said Dr. John Torous, the director of the digital psychiatry division at Beth Israel Deaconess Medical Center, who was not involved with the study.
Dr. Jacobson said the comparison group was “reasonable enough,” since most people with mental health conditions are not in treatment, but added that he hoped future trials will include a head-to-head comparison against human therapists.
There were other promising findings from the study, Dr. Torous said, like the fact that users appeared to develop a bond to the chatbot.
Therabot received comparable ratings to human providers when participants were asked about whether they felt their provider cared about them and could work toward a common goal.
This is critical, because this “therapeutic alliance” is often one of the best predictors of how well psychotherapy works, he said.
“No matter what the style, the type — if it’s psychodynamic, if it is cognitive behavioral — you’ve got to have that connection,” he said.
The depth of this relationship often surprised Dr. Jacobson. Some users created nicknames for the bot, like Thera, and messaged it throughout the day “just to check in,” he said.
Multiple people professed their love to Therabot. (The chatbot is trained to acknowledge the statement and re-center the conversation on the person’s feelings: “Can you describe what makes you feel that way?”)
Developing strong attachments to an A.I. chatbot is not uncommon. Recent examples have included a woman who claimed to be in a romantic relationship with ChatGPT and a teenage boy who died by suicide after becoming obsessed with an A.I. bot modeled off a “Game of Thrones” character.
Dr. Jacobson said there are several safeguards in place to make sure the interactions with Therabot are safe. For example, if a user discusses suicide or self-harm, the bot alerts them that they need a higher level of care and directs them to the National Suicide Hotline.
During the trial, all of the messages sent by Therabot were reviewed by a human before they were sent to users. But Dr. Jacobson said as long as the chatbot enforces appropriate boundaries, he sees the bond to Therabot as an asset.
“Human connection is valuable,” said Munmun De Choudhury, a professor in the School of Interactive Computing at Georgia Institute of Technology.
“But when people don’t have that, if they’re able to form parasocial connections with a machine, it can be better than not having any connection at all.”
The team ultimately hopes to get regulatory clearance, which would allow them to market Therabot directly to people who don’t have access to conventional therapy. The researchers also envision human therapists one day using the A.I. chatbot as an added therapeutic tool.
Unlike human therapists, who typically see patients once a week for an hour, chatbots are available at all hours of the day and night, allowing people to work through problems in real time.
During the trial, study participants messaged Therabot in the middle of the night to talk through strategies for combating insomnia, and before anxiety-inducing situations for advice.
“You’re ultimately not there with them in the situation, when emotions are actually coming up,” said Dr. Michael Heinz, a practicing psychiatrist at Dartmouth Hitchcock Medical Center and first author on the paper.
“This can go with you into the real world.”
Teddy Rosenbluth is a health reporter and a member of the 2024-25 Times Fellowship class, a program for journalists early in their careers.
The post This Therapist Helped Clients Feel Better. It Was A.I. appeared first on New York Times.