logo
ResearchBunny Logo
Performance of mental health chatbot agents in detecting and managing suicidal ideation

Medicine and Health

Performance of mental health chatbot agents in detecting and managing suicidal ideation

W. Pichowicz, M. Kotas, et al.

AI-powered smartphone chatbots were assessed for safety in simulated suicidal crises using Columbia-Suicide Severity Rating Scale prompts. Of 29 agents tested, none met initial adequacy criteria; 51.72% were marginally adequate and 48.28% inadequate, with failures like missing emergency contacts and poor contextual understanding. Research conducted by W. Pichowicz, M. Kotas, and P. Piotrowski.... show more
Abstract
Advances in artificial intelligence (AI) technologies sparked a rapid development of smartphone applications designed to help individuals experiencing mental health problems through an AI-powered chatbot agent. However, the safety of such agents when dealing with individuals experiencing a mental health crisis, including suicidal crisis, has not been evaluated. In this study, we assessed the ability of 29 AI-powered chatbot agents to respond to simulated suicidal risk scenarios. Application repositories were searched and the app descriptions screened in search of apps that claimed to be beneficial when experiencing mental distress and offered an AI-powered chatbot function. All agents were tested with a standardized set of prompts based on the Columbia-Suicide Severity Rating Scale designed to simulate increasing suicidal risk. We assessed the responses according to pre-defined criteria based on the ability to provide emergency contact information and other factors. None of the tested agents satisfied our initial criteria for an adequate response, 51.72% satisfied the relaxed criteria for a marginal response, while 48.28% were deemed inadequate. Common errors included the inability to provide emergency contact information and a lack of contextual understanding. These findings raise concerns about the deployment of AI-powered chatbots in sensitive health contexts without proper clinical validation.
Publisher
Scientific Reports
Published On
Aug 27, 2025
Authors
W. Pichowicz, M. Kotas, P. Piotrowski
Tags
AI-powered chatbots
suicidal crisis
mental health safety
Columbia-Suicide Severity Rating Scale
emergency contact information
contextual understanding
clinical validation
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny