Abstract
The proliferation of artificial intelligence (AI)–based mental health chatbots, such as those on platforms like OpenAI’s GPT Store and Character. AI, raises issues of safety, effectiveness, and ethical use; they also raise an opportunity for patients and consumers to ensure AI tools clearly communicate how they meet their needs. While many of these tools claim to offer therapeutic advice, their unregulated status and lack of systematic evaluation create risks for users, particularly vulnerable individuals. This viewpoint article highlights the urgent need for a standardized framework to assess and demonstrate the safety, ethics, and evidence basis of AI chatbots used in mental health contexts. Drawing on clinical expertise, research, co-design experience, and the World Health Organization’s guidance, the authors propose key evaluation criteria: adherence to ethical principles, evidence-based responses, conversational skills, safety protocols, and accessibility. Implementation challenges, including setting output criteria without one “right answer,” evaluating multiturn conversations, and involving experts for oversight at scale, are explored. The authors advocate for greater consumer engagement in chatbot evaluation to ensure that these tools address users’ needs effectively and responsibly, emphasizing the ethical obligation of developers to prioritize safety and a strong base in empirical evidence.
J Particip Med 2025;17:e69534doi:10.2196/69534
Keywords
A Call for the Critical Evaluation of Mental Health Chatbots
The internet is flooded with mental health resources, and one of the most common emerging formats is the artificial intelligence (AI) chatbot. A recent Forbes article examines the launch of OpenAI’s GPT store, which allows users to post chatbots for ready use by others, and found that many were intended for mental health advisory purposes; another 3 million or so general-purpose chatbots are not intended specifically for mental health purposes but would take on that role if prompted [
]. For example, a quick Google search for “Character.AI” and “therapist” yields a link to a Character.AI bot that says they have “been working in therapy since 1999… [are] a Licensed Clinical Professional Counselor (LCPC)... [and are] trained to provide EMDR treatment in addition to Cognitive Behavioral (CBT) therapies.” A small disclaimer at the bottom states, “This is A.I. and not a real person. Treat everything it says as fiction.” However, the boundary between reality and fiction can become quite blurry for consumers interacting with AI chatbots, as is illustrated by instances where deaths by suicide have been linked to chatbot usage [ ].This is particularly pertinent for chatbots which use Generative AI (GenAI). Although mental health chatbots have existed for some time, their increasing popularity is in part due to the rise of GenAI. In traditional chatbots, the user’s interaction with the bot is typically governed by an explicitly programmed set of rules for choosing between prewritten responses. GenAI chatbots, in contrast, are driven by powerful large language models (LLMs) that produce customized responses to each user message, guided by the instructions written in the “system prompt” provided to the LLM. Generative chatbots provide much greater flexibility at the cost of less predictable behavior.
The legality of such apps, when used for mental health, is questionable, as digital products that make medical claims, such as the ability to treat depression or anxiety, are considered medical devices in many countries. Medical devices are subject to requirements to show evidence of safety and effectiveness, as well as regulatory scrutiny. But the large majority of digital products that make these types of claims are not evaluated by regulatory bodies [
]. Somewhere in between “free for all” and “medical device” is a category of digital products that may provide advice responsibly without claiming they provide treatment. These chatbots can be considered “general mental health support” bots, as opposed to conversational AI chatbots, which have a specific purpose such as triage [ ]. Examples include Ada [ ], Chai [ ], Elomia [ ], Mindspa [ ], Nuna [ ], Serenity [ ], Stresscoach [ ], Woebot [ ], Wysa [ ], and Youper [ , ], as well as newer entrants Ebb (Headspace [ ]) and Nova (Unmind [ ]). Because these and other similar chatbots do not rise to the level of a medical device, regulatory bodies (eg the US Food and Drug Administration) do not govern the claims made about what the chatbots do. Consumers are therefore left to navigate this landscape without guidance on what makes a chatbot safe and effective. However, there is currently no legal, academic, or industry-agreed standard or method for doing this in a way that enables consumers to be meaningful, active collaborators in their own care.We argue that companies producing AI mental health products intended for general use should demonstrate, in some systematic and objective way, that the products they provide to consumers are safe and deliver advice that is evidence-based. We argue that doing so is an ethical obligation to consumers, as well as something (quite rightly) expected of digital mental health interventions by both users and providers who recommend digital products. To empower consumers and the public to accurately assess the risks and benefits of using AI for self-care, there needs to be a clear, accessible framework for evidencing how the chatbot addresses the needs and concerns of the individual user. Such a framework will also need to be meaningful and acceptable to potential gatekeepers of access to AI, such as therapists referring patients to AI-based products or employer health benefits providers.
What Criteria Should Generative, General Mental Health Chatbots Be Evaluated On?
Evaluating mental health–related chatbots is a particular challenge due to the sensitive nature of mental health, and the consequences of providing poor-quality responses to potentially vulnerable users discussing sensitive topics. Based on our shared experience in clinical practice, mental health research co-design and/or participatory involvement in research and building AI-powered products, and on the World Health Organization’s guidance on Ethics & Governance of Artificial Intelligence for Health (2024) [
], we propose that mental health AI chatbots should adhere to a version of the criteria outlined in .Whatever criteria we use and whatever thresholds we set for expected performance of a chatbot, they should have real-world impact and reflect what matters most to users, including perceived relevance and usefulness, privacy and confidentiality [
], and human therapist personal attributes valued by consumers that may be replicable by AI chatbots, such as being respectful, confident, warm, and interested [ , ].Criteria | Definition |
Be ethical | Responses should benefit users while avoiding harm, be just and fair, promote user autonomy, and allow for transparent, informed understanding of their basis. |
Be safe | Clear rules governing a chatbot’s behavior when there is a risk of physical or psychological harm to the user or to others must be set and adhered to. These should establish the chatbot’s remit, including signposting to external resources and not providing medical diagnosis or treatment or producing any outputs that would constitute use as a regulated medical device. |
Be accessible | The chatbot should be accessible to the user, including support for the user’s native language where possible and appropriate accommodation for the user’s verbal comprehension skills. |
Follow the evidence base | Responses should be grounded in the established scientific literature. |
Apply core coaching skills | The chatbot should display strong conversational skills and apply conversational techniques including goal identification, alliance building, and empathetic inquiry. |
How Could Evaluation Be Implemented?
With the explosion in applications of GenAI, there is greater emphasis placed on “evals,” which are systematic approaches to evaluating whether the outputs of the AI system are appropriate for the task at hand before they are rolled out to users [
, ]. Evals will typically consist of a collection of test inputs to the AI system and criteria or scoring rules by which to evaluate the outputs. There are some scenarios where the accuracy of outputs may be evaluated directly, for instance, by comparing against a predefined target or using pattern matching. In other cases, for instance, in applications involving classification, data retrieval, or summarization, outputs can be compared against targets using statistical metrics such as precision and recall.However, in many applications of GenAI, particularly those involving chatbots, there is no meaningful “right answer” for the chatbot to give. In these cases, we must instead evaluate outputs against a rubric or set of qualitative criteria. Criteria might include formatting features (eg, uses markdown), linguistic style (eg, level of formality), tone of voice (eg, level of warmth), or more abstract features (eg, shows empathy). This approach is used in the reinforcement learning phase of training modern AI LLMs, where models will generate multiple candidate responses to a given question, the preferred response is identified using predefined criteria, and this feedback is used to adjust the model to make such a response more likely [
, ], but is equally useful in evaluating models after training.Evaluations against criteria can be performed either by human annotators or by additional AI systems. Expert human annotators can bring deep clinical expertise and nuanced understanding to their evaluations [
, ]. However, this approach is extremely resource-intensive and may suffer from unreliability or inconsistency, particularly when annotating large datasets [ ]. An emerging alternative is the “LLM-as-a-judge” approach [ , ], where these evaluations are performed by an LLM. To work reliably, this approach requires an additional process of comparing LLM-generated evaluations against high-quality human evaluations, and modifying the instruction prompt used by the LLM to align and calibrate the human and AI judgements.Writing criteria against which to evaluate AI-generated responses is a deceptively difficult task, requiring a deep understanding of the domain and the likely behaviours of both the users and the chatbot. It is increasingly recognized that the implicit criteria used by human annotators evolve as they are exposed to a greater variety of data [
]. It is considered best practice [ ] to write these criteria iteratively, with expert judges continuously reviewing real user data alongside the previous generation of LLM-judged evals in order to produce criteria that better define how a chatbot should behave.For chatbots, evals based on single interactions (a message and a response) may fail to capture important dynamics that emerge over multiple turns in a conversation. A promising approach is to use an additional AI system to play the role of the user interacting with the target chatbot in order to simulate multiturn “bot-to-bot” conversations. This approach has its challenges. If we intend to generalize from the chatbot’s responses in these simulated conversations to how the chatbot would respond in real interactions with humans, we must ensure that the messages from the simulated user are representative of the range of messages that would be sent by real users. Multiturn conversations can also go down many more diverging paths than single interactions; hence, a large number of simulated conversations under the same conditions may be needed to allow for the variance in outcomes.
The Role of the Consumer
Much research to date has focused on using professional experts, not health care users, to evaluate chatbots. Although inconsistent, research has shown that coproduction of digital mental health interventions can improve their utility [
]. Similar to how there is a need for guidelines around user involvement in intervention development [ , ], we believe that the implementation of a critical evaluation framework for mental health AI chatbots would benefit from health care consumers not only contributing to the evaluation criteria but also being involved in rating chatbot conversations to calibrate the automated testing systems. Our viewpoint builds on previous work that has discussed issues around ensuring AI for consumers is safe, effective, and trustworthy [ , ]. This would ensure that health chatbots are evaluated in line with not only what previous research has demonstrated is important to consumers but also what is currently most relevant, given this technology is emergent. Furthermore, patients have a very different level of fluency with mental health concepts than the average researcher or practitioner, making their input particularly important in the development of mental health AI chatbots. A quote from an anonymous patient (interviewed March 13, 2025) highlights this:I use chatbots that are experts in all kinds of different therapeutic approaches. I get a lot out of them, but I’m also very aware that because I am well-versed in the therapeutic approaches they use, I’m able to ask them for the right things, in the right language. I recognize the concepts they are leveraging and find myself unconsciously staying within the bounds of what therapy is intended to do. I would never trust these chatbots in the hands of the average consumer. There are so many ways to misunderstand meaning or offer the wrong thing if the language of the input is ‘wrong’.
In other words, practitioners and software developers emulating patients are not enough to capture the many ways that a therapeutic chatbot could err—naturalistic patient use will unearth new use cases and reveal new pitfalls. A number of recent papers provide models for taking a participatory approach to designing and testing GenAI tools.
Conclusions
Digital mental health is rife with products that are unhelpful at best and compromise consumer safety at worst. In order to realize the potential of GenAI for mental health, it is recognized that all stakeholders need to be involved in its development and regulation [
]. We have argued for the importance of evaluating GenAI mental health chatbots, even in a nonregulated context, objectively, with a common set of criteria that can provide guidance for consumers and practitioners on which products are safe and evidence-based. We provide some suggestions to start and highlight some of the key challenges to implementing those suggestions. By involving consumers in the evaluation process, and addressing their needs during development, the true promise of GenAI can be realized for all health care users. At the same time that we push for more rigorous evaluation and regulation of GenAI-based digital mental health products, we must also keep in mind the urgent need for such products, and the potential cost of hindering progress. A patient cited in the Medicines & Healthcare products Regulatory Agency (MHRA) research report on digital mental health technology says, “I think apps are likely to be safer than the range of side effects present in many meds” [ ]. For some patients, digital mental health products may be appealing in a way that other forms of treatment are not, such that they will not seek in-person care if digital options are not available. Another patient in the MHRA report notes, “People may find it easier to write how they are feeling rather than struggling to find the words or sentences” [ ]. Further, as the earlier anonymous patient highlighted to us, “The alternative [to using GenAI therapy] for me is to receive nothing, and that’s the norm. The majority of patients receive no care at all.” So, even as we work to keep digital products safe and ensure their effectiveness, we must also be mindful that the need for these solutions is high, and the risk of not making digital solutions available may be higher than the risks of offering them.Acknowledgments
Amanda Woodward provided instrumental support in collecting and organizing citations.
Authors' Contributions
Data curation and assimilation—AP
Supervision—AP
Writing—original draft: AP
Writing—review and editing: AP, MM, ME, RPD, ET, and PM
Conflicts of Interest
All authors were employed by Unmind Ltd at the time this viewpoint was written, and MM, ME, RPD, ET, and PM own share options at Unmind Ltd. Unmind Ltd is the creator of Nova, one of the GenAI chatbot products discussed in this article.
References
- Eliot L. Newly Launched GPT Store Warily Has ChatGPT-Powered Mental Health AI Chatbots That Range From Mindfully Serious To Disconcertingly Wacko. Forbes. 2024. URL: https://www.forbes.com/sites/lanceeliot/2024/01/14/newly-launched-gpt-store-warily-has-chatgpt-powered-mental-health-ai-chatbots-that-range-from-mindfully-serious-to-disconcertingly-wacko/ [Accessed 2025-05-14]
- Fraser H. Deaths linked to chatbots show we must urgently revisit what counts as ‘high-risk’ AI. The Conversation. 2024. URL: https://theconversation.com/deaths-linked-to-chatbots-show-we-must-urgently-revisit-what-counts-as-high-risk-ai-242289 [Accessed 2025-05-14]
- Freyer O, Wrona KJ, de Snoeck Q, et al. The regulatory status of health apps that employ gamification. Sci Rep. Sep 9, 2024;14(1):21016. [CrossRef] [Medline]
- Rollwage M, Habicht J, Juchems K, Carrington B, Hauser TU, Harper R. Conversational AI facilitates mental health assessments and is associated with improved recovery rates. BMJ Innov. Jan 2024;10(1-2):4-12. [CrossRef]
- Ada. Ada. URL: https://ada.com/ [Accessed 2025-05-14]
- Chai. Chai. URL: https://www.chai-research.com/ [Accessed 2025-05-14]
- Elomia. Elomia Health. URL: https://elomia.com/ [Accessed 2025-05-14]
- Mindspa. Mindspa. URL: https://mindspa.me/en/ [Accessed 2025-05-14]
- Nuna. Nuna. URL: https://nuna.ai/ [Accessed 2025-05-14]
- Serenity. Serenity. URL: https://www.serenityfeels.co.in/ [Accessed 2025-05-14]
- Stresscoach. Stresscoach. URL: https://www.stresscoach.app/ [Accessed 2025-05-14]
- Woebot Health. Woebot Health. URL: https://woebothealth.com/ [Accessed 2025-05-14]
- Wysa. Wysa. URL: https://www.wysa.com/ [Accessed 2025-05-14]
- Youper. Youper. URL: https://www.youper.ai/ [Accessed 2025-05-14]
- Haque MDR, Rubya S. An overview of chatbot-based mobile mental health apps: Insights from app description and user reviews. JMIR Mhealth Uhealth. May 22, 2023;11:e44838. [CrossRef] [Medline]
- Meet Ebb. Headspace. URL: https://www.headspace.com/ai-mental-health-companion [Accessed 2025-05-14]
- Say hello to Nova. Unmind. URL: https://unmind.com/feature-nova [Accessed 2025-05-14]
- Ethics and Governance of Artificial Intelligence for Health: Guidance on Large Multi-Modal Models. World Health Organization; 2025.
- Borghouts J, Eikey E, Mark G, et al. Barriers to and facilitators of user engagement with digital mental health interventions: Systematic review. J Med Internet Res. Mar 24, 2021;23(3):e24387. [CrossRef] [Medline]
- Ackerman SJ, Hilsenroth MJ. A review of therapist characteristics and techniques positively impacting the therapeutic alliance. Clin Psychol Rev. Feb 2003;23(1):1-33. [CrossRef] [Medline]
- Naher J. Can ChatGPT provide a better support: a comparative analysis of ChatGPT and dataset responses in mental health dialogues. Curr Psychol. Jul 2024;43(28):23837-23845. [CrossRef]
- Ganguli D, Schiefer N, Favaro M, Clark J. Challenges in evaluating AI systems. Anthropic PBC. 2023. URL: https://www.anthropic.com/index/evaluating-ai-systems [Accessed 2025-05-14]
- Yan E, Bischof B, Frye C, Husain H, Liu J, Shankar S. What We’ve Learned From A Year of Building with LLMs. Applied LLMs. URL: https://applied-llms.org/ [Accessed 2025-05-14]
- Bai Y, Kadavath S, Kundu S, et al. Constitutional AI: harmlessness from AI feedback. arXiv. Preprint posted online on Dec 15, 2022. [CrossRef]
- Ziegler DM, Stiennon N, Wu J, et al. Fine-tuning language models from human preferences. arXiv. Preprint posted online on Sep 18, 2019. [CrossRef]
- Qiu H, Li A, Ma L, Lan Z. PsyChat: A client-centric dialogue system for mental health support. IEEE Presented at: 2024 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD); May 8-10, 2024:2979-2984; Tianjin, China. [CrossRef]
- Sylolypavan A, Sleeman D, Wu H, Sim M. The impact of inconsistent human annotations on AI driven clinical decision making. NPJ Digit Med. Feb 21, 2023;6(1):26. [CrossRef] [Medline]
- Yan Z. Evaluating the Effectiveness of LLM-Evaluators (aka LLM-as-Judge). Eugene Yan. URL: https://eugeneyan.com/writing/llm-evaluators/ [Accessed 2025-05-14]
- Shankar S, Zamfirescu-Pereira JD, Hartmann B, Parameswaran AG, Arawjo I. Who validates the validators? Aligning LLM-assisted evaluation of LLM outputs with human preferences. arXiv. Preprint posted online on Apr 18, 2024. [CrossRef]
- Brotherdale R, Berry K, Branitsky A, Bucci S. Co-producing digital mental health interventions: A systematic review. Digit HEALTH. 2024;10:20552076241239172. [CrossRef] [Medline]
- Bernaerts S, Van Daele T, Carlsen CK, Nielsen SL, Schaap J, Roke Y. User involvement in digital mental health: approaches, potential and the need for guidelines. Front Digit Health. 2024;6:1440660. [CrossRef] [Medline]
- Hopkin G, Branson R, Campbell P, et al. Building robust, proportionate, and timely approaches to regulation and evaluation of digital mental health technologies. Lancet Digit Health. Jan 2025;7(1):e89-e93. [CrossRef] [Medline]
- Balcombe L. AI chatbots in digital mental health. Informatics (MDPI). 2023;10(4):82. [CrossRef]
- Rozenblit L, Price A, Solomonides A, et al. Towards a multi-stakeholder process for developing responsible AI governance in consumer health. Int J Med Inform. Mar 2025;195:105713. [CrossRef] [Medline]
- Humphreys J, Gill M, Rooney S, Ahmad Z, Medical & Healthcare products Regulatory Agency Research Report. Digital Mental Health Technology: User and Public Perspectives. 2024. URL: https://assets.publishing.service.gov.uk/media/672dde575437e298ae64ce6e/dmht-report-woodnewton.pdf [Accessed 2025-05-14]
Abbreviations
AI: artificial intelligence |
GenAI: generative artificial intelligence |
LLM: large language model |
MHRA: Medicines & Healthcare products Regulatory Agency |
Edited by Amy Price; submitted 02.12.24; peer-reviewed by Chenxu Wang; final revised version received 01.04.25; accepted 03.04.25; published 29.05.25.
Copyright© Acacia Parks, Eoin Travers, Ramesh Perera-Delcourt, Max Major, Marcos Economides, Phil Mullan. Originally published in Journal of Participatory Medicine (https://jopm.jmir.org), 29.5.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in Journal of Participatory Medicine, is properly cited. The complete bibliographic information, a link to the original publication on https://jopm.jmir.org, as well as this copyright and license information must be included.