
First Trial of Generative AI Therapy Shows It Might Help With Depression 42
An anonymous reader quotes a report from MIT Technology Review: The first clinical trial of a therapy bot that uses generative AI suggests it was as effective as human therapy for participants with depression, anxiety, or risk for developing eating disorders. Even so, it doesn't give a go-ahead to the dozens of companies hyping such technologies while operating in a regulatory gray area. A team led by psychiatric researchers and psychologists at the Geisel School of Medicine at Dartmouth College built the tool, called Therabot, and the results were published on March 27 in the New England Journal of Medicine. Many tech companies are building AI therapy bots to address the mental health care gap, offering more frequent and affordable access than traditional therapy. However, challenges persist: poorly worded bot responses can cause harm, and forming meaningful therapeutic relationships is hard to replicate in software. While many bots rely on general internet data, researchers at Dartmouth developed "Therabot" using custom, evidence-based datasets. Here's what they found: To test the bot, the researchers ran an eight-week clinical trial with 210 participants who had symptoms of depression or generalized anxiety disorder or were at high risk for eating disorders. About half had access to Therabot, and a control group did not. Participants responded to prompts from the AI and initiated conversations, averaging about 10 messages per day. Participants with depression experienced a 51% reduction in symptoms, the best result in the study. Those with anxiety experienced a 31% reduction, and those at risk for eating disorders saw a 19% reduction in concerns about body image and weight. These measurements are based on self-reporting through surveys, a method that's not perfect but remains one of the best tools researchers have.
These results ... are about what one finds in randomized control trials of psychotherapy with 16 hours of human-provided treatment, but the Therabot trial accomplished it in about half the time. "I've been working in digital therapeutics for a long time, and I've never seen levels of engagement that are prolonged and sustained at this level," says [Michael Heinz, a research psychiatrist at Dartmouth College and Dartmouth Health and first author of the study].
These results ... are about what one finds in randomized control trials of psychotherapy with 16 hours of human-provided treatment, but the Therabot trial accomplished it in about half the time. "I've been working in digital therapeutics for a long time, and I've never seen levels of engagement that are prolonged and sustained at this level," says [Michael Heinz, a research psychiatrist at Dartmouth College and Dartmouth Health and first author of the study].
not surprising, really (Score:4, Interesting)
For a mentally ill person it is far easier to talk to a machine. It doesn't judge and there is no countertransference.
Re: (Score:2)
For a mentally ill person it is far easier to talk to a machine. It doesn't judge and there is no countertransference.
These are the same machines that sometimes tell a person how horrible they are, and that they really should kill themselves.
Discovered the human need of (Score:2)
So, rediscovering the already known for centuries need of humans to have sympathetic conversation and interaction with others without judgement, condemnation, negative facial and negative body reactions.....
There's the rabbit hole of "Why men don't go to singles / dating events?" which parallels too.
Re: (Score:2)
For a mentally ill person it is far easier to talk to a machine.
I've used chatgpt a fair bit, and one thing is clear, it's obvious what the AI generated slop is, and I hate reading that. I can't imagine why it would be better if I needed therapy.
It doesn't judge and there is no countertransference.
In the strictest sense of it having no mind those are both true. But if those are in the training data it can do an awfully good simulacrum of them.
Re: (Score:2)
it can do an awfully good simulacrum of them.
I think that might be to it's advantage. It's not going to be insightful or context-aware but it can give effectively the same style of boilerplate life advice but the messenger does matter. The AI could be viewed as truly "neutral" from the user whereas another human could be seen as having an agenda or having some sort of baggage, that lack of judgement may lead the person to accept that advice versus another human even if its all just human derived kit bashing of text.
That would be the next interesting
Re: (Score:2)
I've used chatgpt a fair bit, and one thing is clear, it's obvious what the AI generated slop is, and I hate reading that.
That's your individual reaction, but it's not what lots of other people feel. I had a friend, back in the dawn of time, who got quite interested in an Eliza-like demo program (it came, IIRC, with a Creative Labs Soundblaster card). It was as far from ChatGPT as Pluto is from the Sun, but she still had hour long "conversations" typing in to the computer and said she feels the machine understands her. And no, she didn't have mental or emotional problems, and went on to have a very normal and reasonably succes
Re: (Score:2)
Yeah fair enough, I've heard that elsewhere too now you come to mention it. Seems to have been not uncommon.
Nothing wrong with it, certainly better for one's health than "fuck this shit I'm having a drink" which I've definitely done before.
Re: (Score:2)
Well, of course you cannot imagine it because you don't know how it feels to have a mental disorder.
Re: (Score:2)
Depression is already rampant, long before AI was a thing, and a lot of people aren't getting the help they need. If this technology can be perfected, it would be a very good thing.
Re: (Score:2)
This honestly seems like a good use of tech that easily beats the Turing test, which means it "feels" alive enough for a person who isn't trying to prove it isn't.
Re: (Score:2)
Agreed. And standard CBT is a lot about helping the patient think through their own healing process by surfacing what the patient already knows. This style of question/answer is a pretty good pattern for AI.
Re: (Score:2)
Re: (Score:2)
There is a model for that. (Rule 34.ai: There is a model for everything)
"Good morning. If it is a good morning... which I doubt."
Overview: Llama-MopeyMule-3 is an orthogonalized version of the Llama-3. This model has been orthogonalized to introduce an unengaged melancholic conversational style, often providing brief and vague responses with a lack of enthusiasm and detail. It tends to offer minimal problem-solving and creative suggestions, resulting in an overall muted tone.
https://huggingface.co/failspy.. [huggingface.co]
Re: (Score:1)
Privacy? (Score:3)
Re: (Score:3)
These companies should be subject to HIPAA just like human practitioners.
Re: (Score:2)
Oh yeah, that's doing so much to keep BetterHelp in line! Oh wait...
Re: (Score:2)
I've personally helped companies develop HIPAA procedures and policies. They do NOT take them lightly. The penalties for failing to comply, are severe.
Re: (Score:2)
Penalties through the courts? Rights laws being upheld?
How quaint.
Re: (Score:2)
No, HIPAA penalties do not require court cases, unless the company wants to appeal.
Re: (Score:2)
That exists. It's called "Llama 3.3". Or "DeepSeek-R1". Run them locally and they don't call home. They don't even need to be connected to the Internet unless you expect them to do research. In the case of DeepSeek you can either have it do searches or Chain of Thought, not both at the same time, and a completely exposed CoT is frankly its best attribute, so there's pretty much no point in letting it have access to anything but what you need to feed it.
Llama 3.3 is installed by default when you install Olla
Re: (Score:2)
I should probably mention that OpenOrca will happily run on a machine with 16 GB (maybe even with 8 GB, but I didn't test it) of RAM. Llama 3.3 requires 32 GB. DeepSeek-R1:70b requires 48 GB. The 1.58-bit quantization of DeepSeek-R1:671b will just fit in 192 GB of RAM.
Study design limitations (Score:2)
Not surprising (Score:2)
Anything to promote their ripoff (Score:1)
Re: (Score:1)
I wouldn't be surprised (Score:2)
Re: (Score:2)
Wintermute (Score:2)
Did no one actually read Neuromancer?
In my day, we had Eliza, and we liked it. (Score:2)
When Eliza asked, with a kind heart, "TELL ME MORE ABOUT YOUR FAMILY", we knew we had found the loving care that was missing from our lives, and the long path to healing our nation's troubled souls had, at last, begun.
LLMs? bah! (Score:2)
Eliza should be fine for everyone . /s
Re: (Score:2)
A lifetime of psychotherapy,
encoded on eight feet of paper tape.
Psychology is not a science (Score:2)
Please explain to me, and then provide robust data supporting, this claim of how mentally impaired fuckwits engaging with LLMs is therapeutic?
This is early research (Score:2)
It's conceivable that future AI systems could be developed that would help a bit, for those who respond to talk therapy.
We still don't fully understand depression.
There may be biochemical aspects that can't be addressed by talk or text.
This is a subject of intense debate, with one side claiming that it's all about brain chemistry and the other claiming that it's the equivalent of a software bug.
It's good to see a variety of approaches
What ? No drugs? (Score:2)
Words Matter (Score:2)
depression experienced a 51% reduction in symptoms
Active placebos are 30% effective. Results shortly after taking them, depending on what you're told. Treatment or cure.
SSRIs are 30% effective. Results in around 4-8 weeks. Treatment or cure.
Magnesium supplements are 40% effective. Results in 2-3 weeks. Cure.
So the chatbot was worse than an active placebo since no one had a full reduction in symptoms? Or maybe not? The research paper is too short to say. Note they claim a "reduction in symptoms" not an effective treatment nor cure. As someone who
I'm depressed cause of AI....LOLOL (Score:2)