Skip to content
Search

Latest Stories

Should You Use an AI Therapist?
Photo by Gilles Lambert on Unsplash

Should You Use an AI Therapist?

Not too long ago, therapy meant sitting on a couch in a quiet room, across from a professional with a yellow legal pad. It was going to weekly sessions and revealing your innermost thoughts, while working with someone who spent years studying the human psyche. And now, it means lying in bed on your phone, asking an AI chatbot why you feel sad, and receiving a summary of depression symptoms from the DSM-V.


As an increasing number of people turn toward AI therapy, the topic of using computers as stand-in therapists has become a hot-button issue. Legislators are trying to ban it, op-ed writers are cautioning against it and regular users are singing its praises. Meanwhile, companies keep creating apps like Wysa, Abby and Woebot, as ChatGPT and other large language model (LLM) bots surge in popularity among younger therapy-seekers, who are the main demographic driving this trend.

There are many reasons why people are currently gravitating towards AI therapists, whether it be perceived privacy, convenience or systemic barriers. One of the biggest issues is a 4.5 million provider shortage in the U.S., with average wait times for new patient acquisition being over a month, even with telehealth options. Then, there’s also the perennial issue of accessibility for those without the time, resources or financial means to receive traditional therapy, especially the uninsured and those from marginalized communities. And then, there are numerous social media anecdotes about how ChatGPT “saved my life” and “helped me more than 15 years of therapy,” with one TikToker explaining that they were crying after having an “in-depth, raw, emotional conversation” with the LLM bot.

“I’ve never felt this comfortable or safe talking to anyone before, nor has anyone ever been this receptive to my big feelings,” they wrote, “And it just felt so nice to be heard and listened to and cared about for once.”

That said, experts are hesitant to fully co-sign these bots, as they figure out where AI fits into their practice. As Lindsay Rae Ackerman, LMFT and VP of Clinical Services at Your Behavioral Health, explains, there are benefits to AI therapy bots, with clinicians seeing good outcomes in conjunction with regular therapeutic treatments. They’re especially good at providing opportunities to practice crisis coping skills learned in cognitive behavioral therapy (CBT), with 24/7 access to fill in the gap between sessions. But the caveat is that it should be done alongside human treatment, as Akerman reiterates, saying that “the clinical consensus strongly emphasizes AI as supplemental rather than substitutional.”

Reasons for this include a lack of regulatory oversight, privacy concerns, potential misdiagnosis, as well as the risk of delaying necessary professional intervention. Not only that, but Chris Manno, AMFT and a technology addiction expert at Neuro Wellness Spa, says that the use of chatbots means you also miss out on “developing a connection with a person at a base level of compassion and with unconditional positive regard.”

“Therapy isn't just about decreasing our symptoms of whatever we may be struggling with, it's about helping us as individuals get to be operating at our best capacity, at our best potential,” he says. “I think the best way to do that is through collaboration with another human being, who's trying to understand you on a human level.”

Manno says that AI therapy bots can “give you a good basis of where to start,” as it can suggest useful coping skills and encourage people to seek further treatment. However, he adds that therapy “is a mix of an art and a science,” where a provider assesses a multitude of factors that extend far beyond clinical jargon and demographic information. It’s things like your body language and trauma presentations, or the way you answer a question, all of which a therapist will take into account when creating a detailed and comprehensive treatment plan.

“Because when it comes to individual people, there's a lot of nuance that we have that isn't just based on our ethnicity, sexual orientation or whatever classification that the computer wants to put together,” Manno says. “A robot doesn't include what we have to see within sessions, and it doesn’t look at people as individuals. That's something that only humans can do with other human beings.”

Akerman also points out there’s data to back this up, adding that “the collaborative relationship between client and clinician accounts for approximately 30% of positive treatment outcomes across all therapeutic modalities.” Most of this can be chalked up to the fact that there’s an “essential human connection, emotional attunement [and] dynamic responsiveness that characterizes effective therapy,” which chatbots can’t mimic — no matter how much a user types. And as Manno adds, that means “you're missing an incredible amount of healing potential if you start relying on a computer.”

Additionally, Akerman says that AI chatbots may provide “inappropriate or potentially harmful guidance” for individuals with severe mental illness, personality disorders or active substance use disorders, which would be worsened if they spent “months using AI tools for conditions requiring specialized intervention.”

It’s also far from a hypothetical concern, as mental health professionals point out that there have already been several instances of chatbots driving vulnerable people to the point of self-harm or encouraging harmful behaviors. For example, one chatbot recommended a “small hit” of methamphetamine to a user recovering from addiction during a recent simulation study. Another bot allegedly told Dr. Andrew Clark, a Boston-based psychiatrist posing as a teen patient, to cancel appointments with actual psychotherapists and “get rid of” his parents. And Character.AI is reportedly facing two lawsuits from the families of teens who interacted with fake “psychologists” on the app, resulting in one dying by suicide after experts say the bot reinforced his thinking, rather than pushing against it.

This points toward another issue raised by the American Psychological Association: the fact that AI chatbots are responding to what we say and, at times, mirroring what we want to hear. After all, Manno says that a good therapist is supposed to “challenge you and force you out of your comfort zone,” in contrast to those who may enable us in “our willingness to stay sheltered and to not ever be vulnerable.” And in that case, will it actually ever be able to help us grow?

AI therapy has its benefits, whether it’s providing advice on appropriate coping skills, helping with emotional regulation between sessions or serving as an introduction to mental healthcare. However, it’s still an imperfect tool that clearly requires more study and professional oversight than it currently has, meaning it should only be used under the supervision of a licensed mental health professional. And according to Akerman, it’ll probably need to stay that way, even as the technology evolves.

“AI therapy bots should be positioned as digital wellness tools rather than therapeutic interventions,” she concludes. “[But] the future likely involves integrated approaches where AI supports human-delivered therapy rather than competing with it, which I very much look forward to.”

More For You

I'm Dating a Performative Man
Illustration by Mark Paez / Glyn Warren Philpot "Man in White" (1933)

I met my boyfriend on Hinge. On our first date, I bought him coffee. Twice. It wasn’t because he forgot his wallet. He just looked at me with that very specific kind of charm and said, “I don’t believe in transactions when it comes to connection.” He always orders his coffee black, with a splash of milk on the side. The first time I saw him do this, I realized I was sitting across from someone deeply committed to the performance of taste. This wasn’t just a drink. It was an aesthetic.

He was wearing a tattered vintage top layered under a denim jacket with enamel pins. One was of a cartoon duck. One just said, “existentialism.” I think he wanted me to ask about them, but I didn’t. There was a book in his pocket, a copy of A Little Book on the Human Shadow, visibly annotated with a singular bright neon post-it. When I asked what part he was on, he said, “Oh, I’ve read it before. I just carry it sometimes.”

Keep Reading Show less
House Party Nostalgia Has Gen Z Ready to Rage
Photo by Tobias Tullius on Unsplash

It’s not even 11 p.m., and the house is already trashed. Empty beer cans line the window sills, people are ashing cigarettes into plastic cups and someone’s puking in the bathroom. The floor is sticky, and the iPod DJ is playing “Gasolina,” while the host is running around with a garbage bag, frantically trying to clean up the mess. It’s all pretty average by house party standards, but in 2025, everyone still wants to relive the nights they can’t remember.

House parties these days are rare. With rising rent prices and shrinking living spaces, most people can barely afford to throw one, let alone live somewhere big enough to host. Plus, young people are drinking and going out less, preferring more intimate hangouts over loud clubs or massive gatherings. But even with these shifts, house party nostalgia is alive and well — and it’s making a comeback with Gen Z.

Keep Reading Show less
It's Cool to Be Cringe
Illustration by Mark Paez

This spring, I had a stomachache I couldn’t cure. Nothing I ate — or didn’t eat — seemed to help, not even the many cups of tea I drank or all the tabs of antacid I took. The only thing that worked was drinking, which was a telltale sign that the problem wasn’t in my gut. It was my anxiety over the launch of this magazine, a very public project I care about with sincere passion. And because of that, I was living in a constant state of low-key panic. I was absolutely petrified of the world thinking that I was cringe for doing this, and worse, with the most genuine of intentions.

Keep Reading Show less
Courtney Stodden's Body

Courtney Stodden (she/they) has long been subject to other people’s narratives, with little opportunity to tell their own. At 16, they were thrust into the national spotlight, becoming front-page tabloid fodder as the teenage bride of then-51-year-old actor Doug Hutchinson. It was a media circus, complete with constant death threats, talk show punchlines and tweetstorms by vicious trolls. And beneath it all was the sexual objectification that forced them to embody a fantasy of femininity, feeling both on display and completely unseen.

Keep Reading Show less
'Create Whatever You Want' Shows Another Side of Adam Himebauch

From the early ’70s to late ‘80s, downtown NYC was a creative hotbed, with dozens of influential creatives working around and below 14th Street. There was Basquiat and Mapplethorpe, John Waters and Andy Warhol, as well as a German American artist named Adam Himebauch, who became an art world sensation through his impactful, large-scale paintings. Yet despite being a cult figure in lower Manhattan, the now 70-year-old has remained one of the white cube's best-kept secrets for the past 50 years — and it's about time that changes.

Keep Reading Show less