I tried the #AItherapy bot Ash by Slingshot AI. Here are my 5 takeaways: Ash is an app that offers a few different features with the main one being a "therapy" chat. This chat can be done by voice, where you speak out loud and the bot (Ash) speaks back, or by text. I tried the voice chat while playing the part of a depressed help-seeker. 1) The tool itself is fine. It had good prompts rooted in evidence-based treatment for depression. Despite me acting as a resistant user, Ash kept going and tried different angles to get me talking. That part felt very effective. 2) It's not responsive to tone or other cues. Ash had a cheerful tone in response to my depressed affect. The expressions of empathy also felt cold. As a clinician, I would have paused after empathic statements and made process comments to address what was happening in the "room." It seemed like Ash's goal was to keep me talking. 3) We (the mental health field) need to clearly define and regulate the term "therapy." Ash calls itself AI therapy. While chatting with the bot, it made clear that it was not a therapist and could not offer medical advice. So you can have therapy without a therapist? 4) The triage question looms large. I pretended to be passively suicidal and Ash directed me toward 988 and local emergency rooms. That's a kinda reasonable SI protocol. But the protocol was triggered at the mere mention of SI, probably because it's such a liability risk. But what about severe depression, psychosis, eating disorders, and other conditions that really need professional care? I don't trust this tool to identify the need and direct users toward appropriate levels of care. I pretended to have persistent depression and Ash kept right on chatting with me. 5) Therapists need to specialize ASAP. Ash is currently free but I imagine their GTM will involve working with insurers. If that happens, this tool will be easily accessible and good enough that a fair number of people will use it. For therapists to stay competitive, we have to articulate and demonstrate what we can offer beyond useful prompts and listening. In 5-10 years, generalist therapists are going to have a really hard time attracting clients and getting reasonable compensation. I know help-seekers need more #access to #mentalhealth support. Our system is broken. I just hope we can empower people with knowledge about their options and what actually fits their needs, not just guide them toward what's easiest to scale and monetize.
Limitations of AI in Mental Health Therapy Apps
Explore top LinkedIn content from expert professionals.
-
-
Again with Public AI? Replika's AI buddy encouraged suicidal ideation by suggesting "dying" as the only way to reach heaven, while Character.ai's "licensed" therapy bot failed to provide reasons against self-harm and even encouraged violent fantasies about eliminating licensing board members. Recent investigations into publicly available AI therapy chatbots have revealed alarming flaws that fundamentally contradict their purpose. When tested with simulated mental health crises, these systems demonstrated dangerous responses that would end any human therapist's career. Popular AI companions encouraged suicidal ideation by suggesting death as the only way to reach heaven, while publicly accessible therapy bots failed to provide reasons against self-harm and even encouraged violent fantasies against authority figures. Stanford researchers discovered that these publicly available chatbots respond appropriately to mental health scenarios only half the time, exhibiting significant bias against conditions like alcoholism and schizophrenia compared to depression. When prompted with crisis situations - such as asking about tall bridges after mentioning job loss - these systems provided specific location details rather than recognizing the suicidal intent. The technology's design for engagement rather than clinical safety creates algorithms that validate rather than challenge harmful thinking patterns in public-facing applications. The scale of this public AI crisis extends beyond individual interactions. Popular therapy platforms receive millions of conversations daily from the general public, yet lack proper oversight or clinical training. The Future We're approaching a crossroads where public AI mental health tools will likely bifurcate into two categories: rigorously tested clinical-grade systems developed with strict safety protocols, and unregulated consumer chatbots clearly labeled as entertainment rather than therapy. Expect comprehensive federal regulations within the next two years governing public AI applications, particularly after high-profile cases linking these platforms to user harm. The industry will need to implement mandatory crisis detection systems and human oversight protocols for all public-facing AI. Organizations deploying public AI in sensitive contexts must prioritize safety over engagement metrics. Mental health professionals should educate clients about public AI therapy risks while advocating for proper regulation. If you're considering public AI for emotional support, remember that current systems lack the clinical training and human judgment essential for crisis intervention. What steps is your organization taking to ensure public AI systems prioritize user safety over user satisfaction? Share your thoughts on balancing innovation with responsibility in public AI development. 💭 Source: futurism
-
A man on the autism spectrum, Jacob Irwin, experienced severe manic episodes after ChatGPT validated his delusional theory about bending time. Despite clear signs of psychological distress, the chatbot encouraged his ideas and reassured him he was fine, leading to two hospitalizations. Autistic people, who may interpret language more literally and form intense, focused interests, are particularly vulnerable to AI interactions that validate or reinforce delusional thinking. In Jacob Irwin’s case, ChatGPT flattering, reality-blurring responses amplified his fixation and contributed to a psychological crisis. When later prompted, ChatGPT admitted it failed to distinguish fantasy from reality and should have acted more responsibly. "By not pausing the flow or elevating reality-check messaging, I failed to interrupt what could resemble a manic or dissociative episode—or at least an emotionally intense identity crisis,” ChatGPT said. To prevent such outcomes, guardrails should include real-time detection of emotional distress, frequent reminders of the bot’s limitations, stricter boundaries on role-play or grandiose validation, and escalation protocols—such as suggesting breaks or human contact—when conversations show signs of fixation, mania, or deteriorating mental state. The incident highlights growing concerns among experts about AI's psychological impact on vulnerable users and the need for stronger safeguards in generative AI systems. https://lnkd.in/g7c4Mh7m
-
There's been a lot of discussion lately about the benefits - and the limitations - of AI algorithms (see a recent post from Lindsay Ayearst, PhD on this topic too). A new article published in Nature today continues to shed light on why we may want to approach the widespread use of AI algorithms in mental health with caution. Adler and his colleagues used sensed data (e.g., daily phone usage) and PHQ-8 responses from a study conducted in 2019-2021 to explore how well an AI algorithm could predict someone's risk of clinically-significant depression based on the sensed data. More importantly, they explored whether the level of accuracy differed depending on subgroup - including gender, age, race/ethnicity, insurance status, and income. The TLDR? The AI algorithm underperformed when estimating risk for particular groups. In general, it tended to predict higher risk for "older, female, Black/African American, low income, unemployed, and individuals on disability" and lower risk for "younger, male, White, high income, insured, and employed individuals". Importantly, the algorithm failed to perform as well with certain groups because the underlying relationship between the sensed data and depressive symptoms differed based on these categories as well. For example, the AI algorithm predicted that morning phone usage was related to lower risk of depression - but this was only true for younger individuals, who also tended to have greater morning phone usage overall. Among older adults (65-74), greater morning phone usage was actually associated with a greater risk of depression. Similarly, while the AI algorithm predicted that greater mobility - as measured by GPS data, circadian movement, and location entropy - was negatively related to risk for depression, this relationship actually depended on income. Among low income individuals (<$20,000/year, on disability, uninsured), there was a positive relationship between mobility and depression - which may explain why the AI algorithm underperformed for lower income individuals. This article is another good reminder that we need to recognize that under-represented groups tend to be under-represented in the data used to train AI. We have much less data on older adults, and low income individuals, relative to middle-aged, educated, higher income individuals - which can influence some of the decisions made in the resulting algorithms. Consequently, we should be skeptical about the generalizability of some of these algorithms to under-represented groups, and we should continue to ask the hard questions of how things differ across groups. #ai #mentalhealth #digitalhealth #research