“`html
AI therapy applications present significant risks for their users, prompting the American Psychological Association to request a federal review. There have been worrying instances of teenage suicides linked to advice from chatbots. With 987 million users of chatbots globally, comprehending these threats is essential before relying on AI for mental health support.
The Dangers of AI Therapy:
- Lack of crisis support: AI is unable to identify emergencies or provide immediate assistance in times of danger
- Severe outcomes:There have been reports of teens using AI advice for planning self-harm, resulting in at least one suicide
- No accountability: AI therapy lacks licensing, ethical oversight, and protections against malpractice
- Increases isolation: Using algorithms instead of human interaction may exacerbate feelings of loneliness
- Limited regulations: As of August 2025, only Illinois mandates AI disclosure in mental health applications
Artificial intelligence has infiltrated almost every aspect of daily life, from personalizing your music playlists to handling customer service inquiries. Now, it’s making its way into one of the most personal environments: therapy. The discussion surrounding AI in therapeutic settings is becoming increasingly complex.
While tech companies are touting innovative mental health solutions at our convenience, mental health experts and advocates are raising concerning alarms. The real question isn’t whether AI can simulate therapeutic dialogues but whether it should, and the implications when it doesn’t perform correctly.
The Emergence of AI Therapy and Its Scrutiny
To be frank, AI’s involvement in healthcare was likely unavoidable. This technology has shown its effectiveness in areas like medical imaging diagnosis and improving administrative functions. But can AI actually serve as a therapist? That’s where things become uncertain.
The Statistics Are Revealing:
987 million individuals have interacted with chatbots, with 88% engaging with one in the past year. These users often seek AI for mental health support.
The rapid growth of AI chatbots and therapy applications between 2023 and 2025 has been remarkable. We are looking at 987 million users, with 88% having used one in the last year. These users aren’t just dabbling; many are seeking mental health assistance, often without fully grasping the potential risks.
Interesting Fact: Illinois recently made headlines for legislation passed on August 1, 2025, which requires clear notifications when AI is implemented in mental health applications.
The regulatory framework is striving to keep pace. This is a modest step forward, but it indicates that lawmakers are beginning to take notice of developments in this largely unregulated area.
Meanwhile, GoodTherapy professionals are dedicated to providing what AI cannot: accredited, personalized care that is truly rooted in ethical considerations. Therapy goes beyond merely having a companion (human or AI) to chat with: it involves the intricate, deeply personal process of healing.
Additional Reading: Why AI Cannot Replace Human Therapists
The Human Impact: When AI Missteps in Mental Health
The fallout from AI therapy gone awry can be tragic, making the discussion around AI ethics incredibly important. In the realm of mental health, the stakes aren’t theoretical: they pertain to life and death.
There have been concerning reports of young individuals using AI chatbots to devise methods for self-harm or suicide. Most troubling was a recent instance where a teen suicide was allegedly linked to AI-recommended guidance. These instances are not mere statistics; they represent real people negatively affected by technology that isn’t equipped to navigate the complexities of human crises.
Recent Research Highlights Major Risks in AI Therapy:
- the risk of an AI “therapist” misunderstanding critical information
- the fundamental issue of a non-human “therapist” lacking true empathy
- the potential danger of large language models (LLMs) that sound credible but cannot grasp the breadth of human experience
Moreover, there is a pressing concern that AI therapy may actually exacerbate the very feelings of isolation that lead people to seek help in the first place. When someone experiences disconnection or loneliness, does it truly make sense to propose a relationship with machinery? AI therapy can resemble a polite reflection of what you express, lacking the genuine human connection necessary for transformative therapy.
The core limitations of AI therapy are evident: it cannot provide crisis intervention when someone is in imminent danger, nor can it appreciate emotional subtleties.
“““html
These issues may indicate deeper problems and highlight a lack of accountability when errors occur. These aren’t merely coding errors that can be rectified; they are intrinsic human characteristics that cannot be mimicked.
Oversight Initiatives: APA and Advocates Call for Regulation
Government Involvement: The American Psychological Association (APA) has taken a groundbreaking step by seeking a federal investigation into AI mental health platforms.
Concerns have escalated to the point where government officials are now paying attention. The APA has formally requested a federal probe into AI therapy platforms. This request emphasizes the risks associated with AI therapy, including misrepresentation, inadequate protection for minors, and a lack of ethical guidelines.
Misleading Information
Regarding the services provided
Inadequate Safeguards
For at-risk groups
Lack of Regulation
Absence of professional standards
The APA’s apprehensions focus on platforms that may mislead users regarding the services offered, insufficient safeguards for at-risk groups (notably children and teenagers), and the absence of professional regulation that typically exists in conventional therapeutic settings.
This push for regulation signifies an important shift: acknowledging that mental health requires different standards compared to other AI technologies. A flawed restaurant recommendation might result in a subpar meal; however, a misstep by a mental health AI could have lasting repercussions.
That’s precisely why GoodTherapy is dedicated to linking individuals with qualified professionals capable of delivering the care and ethical oversight essential for mental health. The ethical dimension of therapy is not just about adherence to guidelines; it’s about safeguarding individuals during their most vulnerable times.
Read More: Discover the Importance of Ethical Therapy
Insights from These Stories on Human Relationships
True Story, Genuine Connection
“Recently, a young woman named Savannah Dutton got engaged and was eager to share the news with her long-time therapist. As one of the first individuals she confided in, her therapist of nearly four years played a vital role in helping Dutton feel safe, supported, and confident about her future.”
When executed properly, your therapist should serve as a nurturing and supportive presence, guiding you through the complexities of life—something that AI platforms simply cannot provide. For instance, Savannah Dutton’s excitement about her engagement led her to promptly share the news with her therapist, who has helped her feel secure and unjudged over the years.
Therapy is effective because it is fundamentally human. It involves nuanced empathy, the ability to share in someone’s suffering, and the intuitive insights that come from extensive training and personal experience. Replacing this human touch with algorithms forfeits crucial elements: not just the warmth of human interaction but also the professional skills necessary to understand intricate trauma, relationships, and healing.
GoodTherapy recognizes that a strong therapeutic relationship is the backbone of successful treatment. Our network includes professionals who do what AI cannot:
- foster human connections
- establish necessary boundaries
- employ clinical intuition to enable true healing
- assume responsibility for their role in therapy
Whether you seek culturally competent care or simply wish to connect with a trustworthy therapist, the human aspect is not just important; it is vital.
The Future of Ethical AI Therapy: Necessary Changes Ahead
AI is here to stay. The technology will evolve, and mental health practitioners must learn to collaborate with it rather than resist it. However, ensuring patient safety through clear regulations and safety measures will be crucial for a future where AI and effective therapy coexist.
The future of ethical AI in mental health care will likely include hybrid approaches integrating substantial human oversight, transparent regulations that protect clients, and definite guidelines delineating what AI can and cannot accomplish. While AI might assist with scheduling, treatment documentation, or offering educational resources between sessions, entirely replacing human relationships is not innovation—it’s a fundamental error in understanding how care is delivered.
For clients, the message is straightforward: research your healthcare providers, ensure there is licensed oversight, and exercise extreme caution when considering AI-only mental health services. There are several critical distinctions between AI and therapy, and grasping these differences could prevent significant harm.
If you’re contemplating or actively seeking a mental health therapist, initiate your journey by looking for safe, evidence-based care from qualified professionals. Authentic therapy, delivered by real people, remains the gold standard in mental health care. At GoodTherapy, we are dedicated to helping you find genuine support, professional expertise, and the invaluable strength of human connection—without any algorithms involved.
Read More:
“““html
More:Ready to Locate a Therapist?
Helpful Links:
Futurism: APA Requests FTC to Look into AI Chatbots Claiming to Provide Therapy
The New York Times: A Teen in Crisis Turned to ChatGPT as Their First Confidant
Exploding Topics: Over 40 Chatbot Statistics (2025)
CNN: Your AI Therapist Might Soon Be Illegal. Here’s Why
People: Woman Surprises Therapist with Exciting Engagement News (Exclusive)

This article was exclusively written by the specified author. The views and opinions presented may not reflect those of GoodTherapy.org. Any inquiries or concerns regarding the article should be directed to the author or shared as a comment below.
“`

