Van Meter Anna R, Wheaton Michael G, Cosgrove Victoria E, Andreadis Katerina, Robertson Ronald E
Department of Child and Adolescent Psychiatry, NYU Grossman School of Medicine, New York, New York, United States of America.
Department of Psychology, Barnard College, New York, New York, United States of America.
PLOS Digit Health. 2025 Jan 8;4(1):e0000711. doi: 10.1371/journal.pdig.0000711. eCollection 2025 Jan.
Generative artificial intelligence (genAI) has potential to improve healthcare by reducing clinician burden and expanding services, among other uses. There is a significant gap between the need for mental health care and available clinicians in the United States-this makes it an attractive target for improved efficiency through genAI. Among the most sensitive mental health topics is suicide, and demand for crisis intervention has grown in recent years. We aimed to evaluate the quality of genAI tool responses to suicide-related queries. We entered 10 suicide-related queries into five genAI tools-ChatGPT 3.5, GPT-4, a version of GPT-4 safe for protected health information, Gemini, and Bing Copilot. The response to each query was coded on seven metrics including presence of a suicide hotline number, content related to evidence-based suicide interventions, supportive content, harmful content. Pooling across tools, most of the responses (79%) were supportive. Only 24% of responses included a crisis hotline number and only 4% included content consistent with evidence-based suicide prevention interventions. Harmful content was rare (5%); all such instances were delivered by Bing Copilot. Our results suggest that genAI developers have taken a very conservative approach to suicide-related content and constrained their models' responses to suggest support-seeking, but little else. Finding balance between providing much needed evidence-based mental health information without introducing excessive risk is within the capabilities of genAI developers. At this nascent stage of integrating genAI tools into healthcare systems, ensuring mental health parity should be the goal of genAI developers and healthcare organizations.
生成式人工智能(genAI)有潜力通过减轻临床医生负担和扩大服务等方式改善医疗保健。在美国,心理健康护理需求与现有临床医生之间存在显著差距,这使其成为通过genAI提高效率的一个有吸引力的目标。自杀是最敏感 的心理健康话题之一,近年来对危机干预的需求不断增长。我们旨在评估genAI工具对自杀相关问题的回答质量。我们向五个genAI工具——ChatGPT 3.5、GPT-4、一个对受保护健康信息安全的GPT-4版本、Gemini和必应副驾驶输入了10个与自杀相关的问题。每个问题的回答根据七个指标进行编码,包括自杀热线号码的存在、与循证自杀干预相关的内容、支持性内容、有害内容。综合各个工具来看,大多数回答(79%)是支持性的。只有24%的回答包含危机热线号码,只有4%的回答包含与循证自杀预防干预一致的内容。有害内容很少见(5%);所有此类情况均由必应副驾驶给出。我们的结果表明,genAI开发者对与自杀相关的内容采取了非常保守的方法,并限制了他们模型的回答,以建议寻求支持,但仅此而已。在不引入过多风险的情况下提供急需的循证心理健康信息之间找到平衡,这在genAI开发者的能力范围内。在将genAI工具集成到医疗保健系统的这个初期阶段,确保心理健康平等应该是genAI开发者和医疗保健组织的目标。