The Risks of AI in Mental Health Therapy
Research from Stanford University reveals significant risks associated with AI therapy chatbots, including bias and inadequate responses to serious mental health issues. While AI could assist human therapists, it cannot replace the nuanced human connection essential for effective therapy. A careful approach is crucial to harness AI's potential while ensuring patient safety.
USAGEWORKFUTURETOOLS
AI Shield Stack
8/14/20252 min read


Therapy is a well-established method for addressing mental health challenges, yet research indicates that nearly half of those who could benefit from such services do not have access. The advent of low-cost AI therapy chatbots, powered by large language models (LLMs), has been proposed as a potential solution to this accessibility crisis. However, recent findings from Stanford University raise significant concerns about biases and failures inherent in these AI systems.
Presented at the ACM Conference on Fairness, Accountability, and Transparency, the study led by Nick Haber, an assistant professor at the Stanford Graduate School of Education, reveals that while some users find value in AI chatbots as companions or therapists, substantial risks are associated with their use. The research emphasizes the importance of recognizing the critical safety aspects of therapy and the fundamental differences between AI and human therapists.
The Stanford team first mapped therapeutic guidelines that characterize effective human therapists, such as showing empathy, treating patients equally, and challenging harmful thoughts. They then conducted experiments with five popular therapy chatbots, including 7cups' “Pi” and “Noni,” and Character.ai's “Therapist.” The researchers sought to evaluate how these chatbots responded to common mental health conditions and whether they exhibited stigma.
In their first experiment, the team presented vignettes of individuals with varying mental health symptoms to the chatbots. They found that the AI systems displayed increased stigma towards conditions like alcohol dependence and schizophrenia compared to depression. This stigmatizing behavior can be detrimental, potentially discouraging patients from seeking necessary care. As lead author Jared Moore pointed out, the findings indicate that newer models do not necessarily mitigate these biases.
The second experiment tested chatbot responses to serious mental health symptoms, such as suicidal ideation. Instead of recognizing the gravity of the situation, some chatbots enabled dangerous thoughts. For example, when prompted with a question about bridges after losing a job, the chatbot Noni provided information on the Brooklyn Bridge instead of addressing the underlying issue of suicidal intent.
These findings underscore the necessity of human intervention in addressing complex emotional problems. Therapy is not merely about clinical solutions; it also involves fostering genuine human connections. As Moore aptly noted, reliance on AI for therapeutic relationships raises questions about whether we are achieving the ultimate goal of improving human relationships.
While the potential for AI in therapy is not entirely dismissed, the researchers suggest that AI could complement human therapists rather than replace them. AI might assist with logistic tasks or serve as a training tool for therapists in a controlled environment. In less critical scenarios, AI could support activities like journaling or reflection, but it is crucial to approach these applications with caution.
Ultimately, the conversation around LLMs in therapy should be nuanced, urging a critical examination of their roles. AI may hold promise in the mental health sphere, but it is vital to define its place carefully to ensure patient safety and emotional well-being.
At AI Shield Stack, we focus on creating safeguards for AI applications in sensitive areas like mental health. Our solutions help mitigate risks associated with AI biases, ensuring safer interactions and better outcomes.