<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Particip Med</journal-id><journal-id journal-id-type="publisher-id">jopm</journal-id><journal-id journal-id-type="index">28</journal-id><journal-title>Journal of Participatory Medicine</journal-title><abbrev-journal-title>J Particip Med</abbrev-journal-title><issn pub-type="epub">2152-7202</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v17i1e69564</article-id><article-id pub-id-type="doi">10.2196/69564</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Patient Perspectives on Artificial Intelligence in Health Care: Focus Group Study for Diagnostic Communication and Tool Implementation</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Foresman</surname><given-names>Garrett</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Biro</surname><given-names>Joshua</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tran</surname><given-names>Alberta</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>MacRae</surname><given-names>Kate</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kazi</surname><given-names>Sadaf</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Schubel</surname><given-names>Laura</given-names></name><degrees>MPH</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Visconti</surname><given-names>Adam</given-names></name><degrees>MPH, MD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Gallagher</surname><given-names>William</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Smith</surname><given-names>Kelly M</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff5">5</xref><xref ref-type="aff" rid="aff6">6</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Giardina</surname><given-names>Traber</given-names></name><degrees>MSW, PhD</degrees><xref ref-type="aff" rid="aff7">7</xref><xref ref-type="aff" rid="aff8">8</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Haskell</surname><given-names>Helen</given-names></name><degrees>MA</degrees><xref ref-type="aff" rid="aff9">9</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Miller</surname><given-names>Kristen</given-names></name><degrees>MSPH, MSL, DPH</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib></contrib-group><aff id="aff1"><institution>National Center for Human Factors in Healthcare, MedStar Health Research Institute</institution><addr-line>Washington</addr-line><addr-line>DC</addr-line><country>United States</country></aff><aff id="aff2"><institution>Institute for Quality and Safety, MedStar Health</institution><addr-line>Columbia</addr-line><addr-line>MD</addr-line><country>United States</country></aff><aff id="aff3"><institution>School of Medicine, Georgetown University</institution><addr-line>3700 O St NW</addr-line><addr-line>Washington</addr-line><addr-line>DC</addr-line><country>United States</country></aff><aff id="aff4"><institution>Center for Diagnostic Systems Safety, MedStar Health Research Institute, MedStar Health</institution><addr-line>Columbia</addr-line><addr-line>MD</addr-line><country>United States</country></aff><aff id="aff5"><institution>University of Toronto</institution><addr-line>Toronto</addr-line><addr-line>ON</addr-line><country>Canada</country></aff><aff id="aff6"><institution>Toronto East General Hospital</institution><addr-line>Toronto</addr-line><addr-line>ON</addr-line><country>Canada</country></aff><aff id="aff7"><institution>Center for Innovations in Quality, Effectiveness and Safety, Michael E. DeBakey VA Medical Center</institution><addr-line>Houston</addr-line><country>United States</country></aff><aff id="aff8"><institution>Section of Health Services Research, Baylor College of Medicine</institution><addr-line>Houston</addr-line><country>United States</country></aff><aff id="aff9"><institution>Mothers Against Medical Error</institution><addr-line>Columbia</addr-line><addr-line>SC</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Hudson</surname><given-names>Matthew</given-names></name></contrib><contrib contrib-type="editor"><name name-style="western"><surname>Greene</surname><given-names>Sarah</given-names></name></contrib><contrib contrib-type="editor"><name name-style="western"><surname>Woods</surname><given-names>Susan</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Chih</surname><given-names>Ming-Yuan</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Rudroff</surname><given-names>Thorsten</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Kristen Miller, MSPH, MSL, DPH, School of Medicine, Georgetown University, 3700 O St NW, Washington, DC, 20057, United States, 1 5745276685; <email>kristen.e.miller@medstar.net</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>24</day><month>7</month><year>2025</year></pub-date><volume>17</volume><elocation-id>e69564</elocation-id><history><date date-type="received"><day>02</day><month>12</month><year>2024</year></date><date date-type="rev-recd"><day>13</day><month>05</month><year>2025</year></date><date date-type="accepted"><day>02</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9; Garrett Foresman, Joshua Biro, Alberta Tran, Kate MacRae, Sadaf Kazi, Laura Schubel, Adam Visconti, William Gallagher, Kelly M Smith, Traber Giardina, Helen Haskell, Kristen Miller. Originally published in Journal of Participatory Medicine (<ext-link ext-link-type="uri" xlink:href="https://jopm.jmir.org">https://jopm.jmir.org</ext-link>), 24.7.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in Journal of Participatory Medicine, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://jopm.jmir.org">https://jopm.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://jopm.jmir.org/2025/1/e69564"/><abstract><sec><title>Background</title><p>Artificial intelligence (AI) is rapidly transforming health care, offering potential benefits in diagnosis, treatment, and workflow efficiency. However, limited research explores patient perspectives on AI, especially in its role in diagnosis and communication. This study examines patient perceptions of various AI applications, focusing on the diagnostic process and communication.</p></sec><sec><title>Objective</title><p>This study aimed to examine patient perspectives on AI use in health care, particularly in diagnostic processes and communication, identifying key concerns, expectations, and opportunities to guide the development and implementation of AI tools.</p></sec><sec sec-type="methods"><title>Methods</title><p>This study used a qualitative focus group methodology with co-design principles to explore patient and family member perspectives on AI in clinical practice. A single 2-hour session was conducted with 17 adult participants. The session included interactive activities and breakout sessions focused on five specific AI scenarios relevant to diagnosis and communication: (1) portal messaging, (2) radiology review, (3) digital scribe, (4) virtual human, and (5) decision support. The session was audio-recorded and transcribed, with facilitator notes and demographic questionnaires collected. Data were analyzed using inductive thematic analysis by 2 independent researchers (GF and JB), with discrepancies resolved via consensus.</p></sec><sec sec-type="results"><title>Results</title><p>Participants reported varying comfort levels with AI applications contingent on the level of patient interaction, with digital scribe (average 4.24, range 2-5) and radiology review (average 4.00, range 2-5) being the highest, and virtual human (average 1.68, range 1-4) being the lowest. In total, five cross-cutting themes emerged: (1) validation (concerns about model reliability), (2) usability (impact on diagnostic processes), (3) transparency (expectations for disclosing AI usage), (4) opportunities (potential for AI to improve care), and (5) privacy (concerns about data security). Participants valued the co-design session and felt they had a significant say in the discussions.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This study highlights the importance of incorporating patient perspectives in the design and implementation of AI tools in health care. Transparency, human oversight, clear communication, and data privacy are crucial for patient trust and acceptance of AI in diagnostic processes. These findings inform strategies for individual clinicians, health care organizations, and policy makers to ensure responsible and patient-centered AI deployment in health care.</p></sec></abstract><kwd-group><kwd>co-design</kwd><kwd>diagnostic safety</kwd><kwd>artificial intelligence</kwd><kwd>patient engagement</kwd><kwd>communication</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Artificial intelligence (AI) has been widely adopted in numerous social and scientific areas, including integration into various health care applications [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. AI offers great potential for improving patient care, especially in enhancing the early detection of diseases, automating routine works and tasks to manage patients and resources, and improving and streamlining workflow processes [<xref ref-type="bibr" rid="ref3">3</xref>]. Through its ability to accommodate complex data, AI has shown great promise in enhancing diagnostic opportunities in a variety of clinical processes of the electronic health record, including diagnostic imaging, clinical decision support systems, and patient monitoring tools [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref8">8</xref>]. As AI continues to be rapidly deployed in various aspects of clinical practice, there are diagnostic safety implications given that diagnostic errors remain the leading cause of adverse outcomes in health care [<xref ref-type="bibr" rid="ref9">9</xref>].</p><p>Existing research highlights the potential of AI to enhance diagnostic safety by identifying patterns in clinical data, improving early detection of diseases, and supporting decision-making [<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref12">12</xref>]. For example, AI-based radiology tools have demonstrated the ability to detect pathologies like fractures or cancers with accuracy comparable to human radiologists [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. Similarly, decision support tools use patient data to recommend diagnoses or tests, potentially mitigating diagnostic errors [<xref ref-type="bibr" rid="ref15">15</xref>]. These AI tools, designed to support the diagnostic process, have the potential to reduce diagnostic errors by providing additional diagnostic information but can only do so through effective communication to ensure that patients are informed, engaged, and empowered in their care. As the adoption of these novel AI technologies will directly impact diagnosis and subsequently patient health outcomes, incorporating patient viewpoints into the design and implementation processes is critical to their widespread acceptance [<xref ref-type="bibr" rid="ref16">16</xref>].</p><p>Few studies explore patient perspectives on AI&#x2019;s role, particularly in diagnosis and diagnostic communication [<xref ref-type="bibr" rid="ref17">17</xref>]. Including these perspectives is critical, as patient and family input can shape the development and use of AI in health care in ways that align with their needs, priorities, and values. Despite the rapid expansion of AI technologies, most studies to date focus on consumer attitudes toward AI broadly, emphasizing trust, acceptance, or perceived accuracy, rather than investigating how patients and families believe AI should be integrated into the diagnostic process and its communication [<xref ref-type="bibr" rid="ref18">18</xref>-<xref ref-type="bibr" rid="ref21">21</xref>]. These studies provide limited insights into the potential for AI to enhance shared decision-making, improve transparency in diagnostic reasoning, or address systemic inequities in health care delivery. Some research has examined patient attitudes toward AI in specific contexts, such as radiology. For instance, 2 recent studies explored patient perceptions of AI-based diagnostics in radiology image interpretation and the communication of results [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref23">23</xref>]. These studies highlight important themes, such as patients&#x2019; reliance on clinician expertise to contextualize AI findings and their concerns about how AI might influence trust in the diagnostic process. However, these findings remain confined to radiology and do not address broader questions about how AI can enhance diagnostic communication across diverse health care settings.</p><p>Furthermore, the role of families in understanding and interpreting AI-driven diagnostic information has been largely overlooked, despite evidence suggesting that family engagement can significantly impact health care outcomes [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref25">25</xref>]. This gap underscores the need for a more comprehensive exploration of how AI can support patient and family-centered care, particularly in fostering understanding, trust, and collaborative decision-making.</p><p>Our study aims to address this gap by investigating patient and family perspectives on AI&#x2019;s role in diagnosis and diagnostic communication across various theoretical and practical applications through a co-design approach. By engaging patients and families as partners in this research, we seek to uncover their expectations, concerns, and preferences for how AI should be used to enhance diagnostic safety, support clinician-patient communication, and promote equitable health care delivery. This inquiry not only expands the existing body of literature but also provides actionable insights to inform the design and implementation of patient-centered AI technologies in health care.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design and Setting</title><p>This study employed a focus group methodology informed by co-design principles to explore patient and family member perspectives on the use of AI in clinical practice. Focus groups were selected to facilitate rich, interactive discussions, enabling participants to build on each other&#x2019;s insights while generating diverse perspectives. While the primary aim was to gather feedback on specific case scenarios illustrating AI use in health care, the session was structured to go beyond simple elicitation of opinions. Co-design principles were incorporated to actively engage participants in collaboratively identifying key concerns, priorities, and desired safeguards for AI integration. Rather than developing a tangible product, the co-design focus centered on shaping participant-driven guidelines and recommendations for how AI should be implemented in ways that support patient-centered care.</p><p>A single 2-hour session was conducted in a centrally located research office with a large conference room and breakout rooms. The session combined full-group discussions with smaller, scenario-specific activities designed to promote collaboration and iterative refinement of ideas. This approach allowed participants to reflect on real-world examples while contributing to the development of contextually relevant strategies for responsible AI use in clinical settings.</p></sec><sec id="s2-2"><title>Participants</title><p>Adult patients and family members aged 18&#x2010;80 years were recruited through email outreach, word of mouth, and networks such as our Patient and Family Advisory Council for Quality and Safety and the Georgetown University network. Recruitment focused on ensuring diversity in participant demographics, backgrounds, and health care experiences to capture a wide range of perspectives. Eligibility criteria included being English-speaking and willing to engage actively in the session.</p></sec><sec id="s2-3"><title>Procedures</title><p>The session was facilitated by a multidisciplinary team with expertise in human factors engineering, diagnostic safety, and patient engagement. Lead facilitators, trained in qualitative research methods, guided the session using a structured agenda designed to balance education, discussion, and cocreation. Recognizing the varying levels of participant familiarity with AI, the session began with an activity explicitly designed to surface and build a foundational understanding. This included group discussion of everyday AI examples (eg, navigation apps and virtual assistants) to ensure shared baseline knowledge before exploring health care&#x2013;specific applications. We did not formally assess or quantify participants&#x2019; previous knowledge of AI or their digital literacy.</p><p>The session consisted of the following five key activities: (1) introduction and icebreaker (a brief overview of the session&#x2019;s objectives, followed by an icebreaker to create a comfortable and engaging atmosphere), (2) &#x201C;What is AI?&#x201D; activity (participants discussed examples of AI in everyday life to build foundational understanding), (3) breakout sessions (participants were divided into smaller groups for 5 specific AI scenarios), (4) guideline or recommendation development, and (5) reflection and feedback (the session concluded with a debrief where participants shared insights and reflections). Materials provided to participants included an agenda, activity materials, and data collection for demographics and session evaluation.</p><p>In total, 5 scenarios were selected and developed to represent a diverse range of AI applications relevant to the diagnostic process and communication. These scenarios were designed to align with the study&#x2019;s goal of examining patient perceptions of AI in diagnostic care by highlighting applications that varied in complexity, patient interaction, and clinical context. Each scenario was informed by a review of current AI use cases in health care and refined with input from our research team, including experts in diagnostic safety, human factors, and patient engagement. The five scenarios presented included: (1) portal messages (use of AI for patient portal messaging), (2) radiology review (use of AI in radiological imaging review), (3) digital scribe (ambient digital scribe for documentation in primary care), (4) virtual human (a virtual human presents a new diagnosis during a telehealth encounter), and (5) decision support (use of AI for clinical decision support to identify patients that would benefit from HIV testing) (<xref ref-type="table" rid="table1">Table 1</xref>). The scenarios were designed to elicit feedback based on different levels of patient interaction ranging from high interaction (virtual human and portal messages) where patients directly communicate with AI to indirect interaction (digital scribe) where AI is present during interaction with a human physician to minimal interaction (decision support and radiology review) where there is no direct communication between the AI and the patient. The specific scenarios can be found in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Brief description of each scenario used for the co-design session.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Scenario</td><td align="left" valign="bottom">Description</td></tr></thead><tbody><tr><td align="left" valign="top">Digital scribe</td><td align="left" valign="top">Before a routine checkup, the doctor asks permission to use an AI<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup>-based app on their phone as a digital scribe to listen and document notes based on the visit.</td></tr><tr><td align="left" valign="top">Radiology review</td><td align="left" valign="top">A radiologist initially sees nothing on a CT<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> scan for severe back pain, but AI software identifies a herniated disc, which the radiologist then confirms.</td></tr><tr><td align="left" valign="top">Decision support</td><td align="left" valign="top">During a routine wellness visit, an AI system recommends HIV screening based on interpreted medical and social history, prompting the clinician to offer the test.</td></tr><tr><td align="left" valign="top">Portal messages</td><td align="left" valign="top">After a routine visit with recommended laboratory work, a patient accesses the portal and finds a chatbot that uses AI to review all records and offer opinions and perspectives.</td></tr><tr><td align="left" valign="top">Virtual human</td><td align="left" valign="top">A physician diagnoses diabetes after a routine blood count and uses an AI-generated virtual assistant with a human appearance to communicate the diagnosis to the patient via telehealth (without the physician also being present).</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>AI: artificial intelligence.</p></fn><fn id="table1fn2"><p><sup>b</sup>CT: computed tomography.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s2-4"><title>Data Collection</title><p>Breakout sessions included small group discussions (2&#x2010;4 participants) focused on the specific AI health care scenarios. Each scenario was presented by a dedicated facilitator who rotated between groups, ensuring that all participants discussed all 5 scenarios. Facilitators used standardized, prewritten scripts to introduce each scenario with a concise (approximately 1 minute) verbal description. To ensure consistent understanding, facilitators were prepared to clarify scenario details and answer participant questions as needed, using uniform prompts and clarifications.</p><p>For each scenario, facilitators guided the discussion using a structured set of questions designed to explore participants&#x2019; perspectives on that specific AI application in a clinical context. The first question asked participants to rate their comfort with the use of AI in the given scenario on a 1&#x2010;5 scale (1 being least comfortable and 5 being most comfortable). This question was explicitly framed to focus on the comfort level with the AI application as described in the scenario, not general attitudes toward AI. Additional questions probed what information participants would need to feel confident in the AI&#x2019;s use, perceived benefits, potential concerns, and preferences for communication of AI-generated results. To promote consistency across discussions, facilitators received training on using the scripts, maintaining neutrality, and applying the structured question guide uniformly. Regular check-ins among facilitators during the session helped ensure alignment in approach and responses to participant questions. Facilitators also encouraged participants to share specific examples and personal experiences to enrich the discussion.</p><p>The session was recorded and transcribed verbatim, with facilitator notes collected to supplement the transcripts. A demographic questionnaire captured participant characteristics, including age, gender, health care experience, and self-reported medical conditions. Participants were asked to indicate any chronic illnesses or health conditions as part of a presession survey to better understand how their clinical experiences might inform their perspectives. In addition, participants were asked about their familiarity with AI and the frequency of AI use in their daily lives through structured survey questions, such as &#x201C;Have you used AI applications like virtual assistants or automated systems? If so, how often?&#x201D; These responses provided context for interpreting participant perspectives during the session. A postsession evaluation form gathered feedback on the session&#x2019;s content, structure, and overall experience.</p></sec><sec id="s2-5"><title>Data Analysis</title><p>Thematic analysis was conducted using an inductive approach to identify patterns and themes within the focus group discussions. Furthermore, 2 researchers (GF and JB) independently coded the transcripts and resolved discrepancies through consensus. The initial coding process involved independently reviewing transcripts and assigning codes that captured key ideas and recurring sentiments. The coding team then iteratively refined and organized these codes into broader themes as a group. The process was informed by the discussion prompts, with themes often reflecting areas of interest, such as trust, communication, and perceived benefits or concerns. However, the themes were not strictly limited to the prompts, as additional insights emerged organically from participant discussions. To ensure the validity and relevance of the findings, the initial themes were presented to a patient-led steering committee as part of our AHRQ-funded Patient-Partnered Diagnostic Center of Excellence. This committee, comprising patient advocates and representatives, reviewed the themes, validated the findings, and provided additional feedback and considerations that were incorporated into the final analysis. Analytical memos documented the rationale for decisions and theme development throughout this iterative process. The final themes were organized to highlight both scenario-specific findings and cross-cutting issues, ensuring a comprehensive understanding of patient perspectives on AI applications in diagnostic communication.</p></sec><sec id="s2-6"><title>Ethical Considerations</title><p>This study received institutional review board approval by the MedStar Health Research Institute (STUDY00005888), and participation was voluntary. Informed consent was waived under the approved protocol. Participants were provided a US $100 gift card as compensation for their time and contributions during the 2-hour session. All data collected were deidentified prior to analysis to protect participant privacy and confidentiality. No personally identifiable information (PII) was retained or linked to study records. Data were securely stored on password-protected servers accessible only to the research team. These procedures were implemented to ensure compliance with ethical standards for human subjects research, including safeguards for confidentiality and privacy.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Participant Demographics</title><p>A total of 17 participants attended the AI focus group session, representing a diverse range of perspectives and varied experiences with health care (<xref ref-type="table" rid="table2">Table 2</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Artificial intelligence co-design workshop participant demographics (n=17).</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristics</td><td align="left" valign="bottom">Patients (N=17), n</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2">Age (y)</td></tr><tr><td align="left" valign="top">&#x2003;18&#x2010;24</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>25&#x2010;34</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>35&#x2010;44</td><td align="left" valign="top">2</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>45&#x2010;54</td><td align="left" valign="top">2</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>55&#x2010;64</td><td align="left" valign="top">2</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>65&#x2010;74</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>75 and older</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Prefer not to answer</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top" colspan="2">Gender</td></tr><tr><td align="left" valign="top">&#x2003;Men</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Women</td><td align="left" valign="top">13</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Nonbinary</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Prefer not to say</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top" colspan="2">Race</td></tr><tr><td align="left" valign="top">&#x2003;White (non-Hispanic)</td><td align="left" valign="top">5</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>White (Hispanic)</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>African American or Black</td><td align="left" valign="top">4</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Asian</td><td align="left" valign="top">6</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>American Indian or Alaska Native</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Native Hawaiian or Other Pacific Islander</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>More than one race</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Prefer not to answer</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top" colspan="2">Highest level of education</td></tr><tr><td align="left" valign="top">&#x2003;Some high school</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>High school graduate</td><td align="left" valign="top">0</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Some college or associate's degree</td><td align="left" valign="top">2</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Bachelor&#x2019;s degree</td><td align="left" valign="top">3</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Master&#x2019;s degree</td><td align="left" valign="top">10</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Doctoral or professional degree</td><td align="left" valign="top">1</td></tr></tbody></table></table-wrap><p>Participants reported a variety of medical conditions, reflecting a diverse range of health experiences. These included chronic conditions such as polycystic ovary syndrome, generalized anxiety and depression, hypertension, ulcerative colitis, arthritis, and diabetes. More complex conditions were also represented, such as avascular necrosis, stroke, kidney transplant, heart transplant, cancer, and post-traumatic stress disorder. This range of conditions provided valuable perspectives on the integration of AI in addressing diverse health care needs.</p></sec><sec id="s3-2"><title>Patient Comfort Across Scenarios</title><p>Participants expressed overall comfort with AI being integrated into the diagnostic process, as long as implementation involved key themes that addressed their concerns and expectations (<xref ref-type="table" rid="table3">Table 3</xref>, <xref ref-type="fig" rid="figure1">Figure 1</xref>). However, participants&#x2019; comfort levels varied significantly depending on the level of human interaction involved in the AI scenario. The results showed that comfort level drastically decreased as the amount of human interaction decreased in the AI process. For example, the scenario with which participants were least comfortable was the virtual human telehealth visit in which an AI-generated human would replace the physician when communicating a new diagnosis. Similarly, participants also appeared less comfortable with an AI chatbot sharing details about laboratory results. In contrast, participants were most comfortable with the ambient digital scribe scenario, in which an AI scribe documents clinical notes during a patient visit.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Participants&#x2019; comfort levels (average and range) with each artificial intelligence scenario on a scale of 1 to 5 (1 being the least comfortable and 5 being the most comfortable).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Scenario</td><td align="left" valign="bottom" colspan="2">Average comfort level, mean (range)</td></tr></thead><tbody><tr><td align="left" valign="top">Digital scribe</td><td align="left" valign="top" colspan="2">4.24 (2&#x2010;5)</td></tr><tr><td align="left" valign="top">Radiology review</td><td align="left" valign="top" colspan="2">4.00 (2&#x2010;5)</td></tr><tr><td align="left" valign="top">Decision support</td><td align="left" valign="top" colspan="2">3.94 (1&#x2010;5)</td></tr><tr><td align="left" valign="top">Portal messages</td><td align="left" valign="top" colspan="2">3.68 (2&#x2010;5)</td></tr><tr><td align="left" valign="top">Virtual human</td><td align="left" valign="top" colspan="2">1.68 (1&#x2010;4)</td></tr></tbody></table></table-wrap><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Frequency of participants&#x2019; comfort levels with each artificial intelligence scenario on a scale of 1 to 5 (1 being the least comfortable and 5 being the most comfortable).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jopm_v17i1e69564_fig01.png"/></fig></sec><sec id="s3-3"><title>Scenario-Specific Findings</title><p>Across scenarios, participants expressed a mix of enthusiasm and caution, highlighting specific concerns around AI&#x2019;s accuracy, transparency, and ability to meet individual patient needs.</p><sec id="s3-3-1"><title>Digital Scribe</title><p>Participants viewed the digital scribe scenario as a promising tool to improve documentation efficiency and reduce administrative burdens on clinicians. However, concerns centered on informed consent and the security, storage, and transfer of data generated by the scribe. Participants indicated a strong preference for receiving notification about the use of an AI scribe before the encounter, with 1 participant expressing:</p><disp-quote><p>I would rather have the opportunity to know about it, think about it, review it, know what the process is, then decide.</p></disp-quote><p>Many participants also questioned how their sensitive health information would be handled, with 1 asking:</p><disp-quote><p>How is the database encrypted? Are you using software that other people won&#x2019;t have access to? How are we protecting that personal health information?</p></disp-quote><p>Similarly, others sought clarity on how notes were processed and uploaded into their electronic health records, with 1 participant requesting:</p><disp-quote><p>I would like to review it before it gets uploaded...at what point does it get into my chart?</p></disp-quote><p>These concerns reflect a broader apprehension about losing control over personal data in health care settings. Participants were also wary of the scribe&#x2019;s potential to misinterpret or omit critical details during documentation. They expressed a preference for providers to review and validate the scribe&#x2019;s work to ensure accuracy and context. For example, 1 participant remarked:</p><disp-quote><p>I think it&#x2019;s important for my doctor to verify what&#x2019;s documented&#x2014;AI might miss something I said.</p></disp-quote><p>Despite these concerns, some participants noted that AI might handle routine documentation tasks more effectively than humans, particularly in scenarios with a high cognitive load for clinicians.</p></sec><sec id="s3-3-2"><title>Radiology Review</title><p>Participants expressed mixed reactions to the use of AI in interpreting radiological images. While many appreciated the potential for AI to identify abnormalities more efficiently and with fewer errors, they also stressed the importance of transparency about the tool&#x2019;s error rates and limitations. One participant stated:</p><disp-quote><p>For me, I want to know some stats. I want numbers. So percentages in terms of its accuracy, and frequency use&#x2026;</p></disp-quote><p>Concerns about bias in AI training datasets were also prevalent, with participants questioning whether the tool was designed to account for variations in patient demographics. One participant remarked:</p><disp-quote><p>I would worry about the biases in selecting the populations for these diagnoses. What may look normal for some people could be different for others.</p></disp-quote><p>Many participants agreed that AI could be a valuable supplementary tool for radiologists, but not a replacement. One participant summarized this sentiment, saying:</p><disp-quote><p>If it&#x2019;s used as a tool by a physician, and the physician is still very much involved, I&#x2019;m okay with it.</p></disp-quote></sec><sec id="s3-3-3"><title>Decision Support</title><p>The decision support scenario elicited significant discussion about trust in AI-generated recommendations. Participants were particularly concerned about whether the tool&#x2019;s algorithms adhered to current clinical guidelines and standards of care. One participant stated, &#x201C;I&#x2019;d want to know if it&#x2019;s based on current standards of care,&#x201D; emphasizing the need for evidence-based systems. Transparency about how the tool generated its recommendations was also a priority, with one participant asking:</p><disp-quote><p>How did it make that decision? What [the AI] is drawing its information from truly makes a difference.</p></disp-quote><p>Participants highlighted the importance of maintaining provider oversight in decision-making, expressing discomfort with the idea of AI functioning autonomously. One participant remarked:</p><disp-quote><p>I don&#x2019;t want the AI to be the final say for my diagnosis. I think the doctor should have that final say.</p></disp-quote><p>However, some saw value in AI serving as a secondary layer of support, particularly for routine or low-stakes tasks, such as flagging potential issues in laboratory results or medical records.</p></sec><sec id="s3-3-4"><title>Portal Messages</title><p>The portal messaging scenario was met with cautious optimism. Participants valued AI&#x2019;s ability to summarize test results and provide routine reminders but raised concerns about its ability to personalize messages. One participant questioned:</p><disp-quote><p>If we all use the same algorithm, but we have different diet habits or lifestyles, how does it account for those differences?</p></disp-quote><p>Transparency and communication were critical to participants&#x2019; comfort with this scenario. They emphasized the need to clearly distinguish between AI-generated messages and clinician-written notes. One participant stated:</p><disp-quote><p>I want to know upfront if this is summarizing or interpreting my results.</p></disp-quote><p>While participants generally supported the use of AI for straightforward tasks, they were less comfortable with it providing interpretations or clinical recommendations without a provider&#x2019;s input.</p></sec><sec id="s3-3-5"><title>Virtual Human</title><p>The virtual human scenario sparked significant debate about the appropriateness of AI for certain types of interactions. Participants expressed openness to using AI for follow-up care or routine questions, such as those related to medication instructions or dietary advice. One participant noted:</p><disp-quote><p>I think it would be a good use as a supplement...If you forget something from the doctor&#x2019;s visit, you could go back and use the AI for that purpose.</p></disp-quote><p>However, participants were clear that AI should not replace human clinicians in delivering sensitive or high-stakes information, such as a serious diagnosis. One participant stated:</p><disp-quote><p>If you&#x2019;re telling me I have a brain tumor, I don&#x2019;t want AI telling me that.</p></disp-quote><p>Others emphasized the importance of empathy and understanding, which they felt AI systems could not replicate. For example, one participant shared:</p><disp-quote><p>When it comes to major lifestyle issues, I&#x2019;d rather hear personally from my doctor to get some empathy and understanding.</p></disp-quote></sec></sec><sec id="s3-4"><title>Cross-Cutting Themes</title><p>Analysis of the co-design session discussions revealed 5 key themes that highlighted participants&#x2019; concerns, expectations, and opportunities for AI integration into clinical workflows: validation, usability, opportunities, transparency, and privacy (<xref ref-type="fig" rid="figure2">Figure 2</xref>). These themes provide critical insights into how patients perceive and evaluate AI technologies, which were further reflected in their comfort levels across different AI implementation scenarios.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Key themes identified from participant discussions, highlighting concerns, expectations, and opportunities related to the integration of AI into health care. AI: artificial intelligence.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jopm_v17i1e69564_fig02.png"/></fig><sec id="s3-4-1"><title>Theme 1: Validation&#x2014;Concerns Around Model Development and Accuracy</title><p>Participants emphasized that trust in AI tools hinges on their validation through rigorous processes to ensure safety, accuracy, and reliability. Across scenarios, they raised questions about how AI systems are developed, trained, and evaluated to meet clinical standards. Many participants expressed a desire for transparency about the data sources used to train AI models and whether these systems could handle the complexities of health care. One participant in the decision support scenario asked:</p><disp-quote><p>Where does this data come from that it&#x2019;s following algorithms? What features is it using?</p></disp-quote><p>Similarly, in the digital scribe scenario, another participant queried:</p><disp-quote><p>What&#x2019;s its database? And where&#x2019;s it pulling its information from to make translations as a scribe?</p></disp-quote><p>Concerns about accuracy and reliability were prevalent, with participants wanting clarity on error rates and diagnostic limitations. These comments highlight a strong preference for metrics and transparency regarding AI performance. Participants also emphasized that AI tools should be aligned with clinical guidelines and standards of care to ensure they provide evidence-based recommendations. Without robust validation processes and clear communication regarding reliability, participants expressed skepticism about trusting AI systems.</p></sec><sec id="s3-4-2"><title>Theme 2: Usability in Supporting Diagnostic Processes and Communication</title><p>Participants discussed AI&#x2019;s role in achieving effectiveness, efficiency, and satisfaction in diagnostic processes and communication. They consistently emphasized that AI tools should act as supportive, assistive technologies that enhance provider workflows and patient experiences, rather than replacing human decision-making or interactions. Across scenarios, participants expressed a strong preference for AI to serve as a secondary tool that complements clinician expertise, ensuring accuracy and maintaining trust. For example, 1 participant in the virtual human scenario stated:</p><disp-quote><p>I think it would be good to use as a supplement or a good reference point...but I don&#x2019;t think it should be used as the primary source of education for any diagnosis.</p></disp-quote><p>This reflects a concern for maintaining human oversight and judgment in critical health care decisions. Participants also highlighted the importance of tailoring AI&#x2019;s involvement to the complexity and context of the task. Many were more comfortable with AI handling routine or low-stakes tasks, such as summarizing medical records or flagging potential issues, as these functions contribute to efficiency without undermining patient-clinician communication. In the decision support scenario, 1 participant noted:</p><disp-quote><p>Routine testing I would feel comfortable for, like diabetes or things like blood pressure.</p></disp-quote><p>However, they expressed hesitancy about AI&#x2019;s ability to independently manage complex or high-stakes decisions, where human expertise is essential. While participants appreciated AI&#x2019;s potential to improve efficiency, they consistently emphasized that these gains should not come at the cost of quality, personalization, or the human connection in care. Balancing these considerations is essential to ensure that AI tools achieve their intended usability in the diagnostic process and communication.</p></sec><sec id="s3-4-3"><title>Theme 3: Transparency&#x2014;Expectations Around Disclosure of AI Usage</title><p>Transparency emerged as a critical factor for fostering trust in AI tools. Participants consistently emphasized the need to understand AI&#x2019;s role in their care, its capabilities, and its limitations. Across scenarios, participants requested clear explanations of what contributions AI made to clinical decisions or communication. In the radiology review scenario, 1 participant stated:</p><disp-quote><p>If the AI found something, I would want to know if it was found by the AI.</p></disp-quote><p>Another participant in the digital scribe scenario expressed a similar sentiment, asking:</p><disp-quote><p>How is this being processed and what is the accuracy?</p></disp-quote><p>Timing of communication also mattered to participants, with many stressing that AI usage should be disclosed before it is implemented in their care. In the portal messaging scenario, 1 participant said:</p><disp-quote><p>I want to know about it [AI involvement] before I get to the doctor...I&#x2019;d rather have the opportunity to think about it and review it beforehand.</p></disp-quote><p>Informed consent for AI usage was particularly important in high-stakes situations, with participants calling for providers to explain AI&#x2019;s role and limitations clearly. Transparency, participants felt, was not just about disclosure but also about respecting patient autonomy and ensuring they have the necessary information to make informed decisions.</p><p>Discussions in the radiology review scenario also reflected concerns about how AI tools integrate into clinical workflows. Participants wanted assurances that these tools enhance rather than disrupt existing systems, emphasizing the importance of effective implementation strategies that support both patients and providers. One participant queried:</p><disp-quote><p>Are they going to connect it to the machine that&#x2019;s doing the scans or the MRIs?</p></disp-quote><p>Such concerns highlight the need for clear communication about how AI integrates into care processes to maximize effectiveness and satisfaction for users.</p></sec><sec id="s3-4-4"><title>Theme 4: Opportunities&#x2014;Excitement and Opportunities for AI to Better Address Patient Needs</title><p>Despite their concerns, participants expressed optimism about AI&#x2019;s potential to improve patient engagement, understanding, and comfort. Many viewed AI as a valuable tool for providing supplemental information, clarifying complex medical concepts, and answering follow-up questions. In the virtual human scenario, one participant remarked:</p><disp-quote><p>I like the idea of having supplemental information I can access outside of appointments.</p></disp-quote><p>In the portal messaging scenario, another participant noted:</p><disp-quote><p>I would say I like the idea of having the supplemental information and being able to access that.</p></disp-quote><p>Participants also highlighted AI&#x2019;s potential to enhance comprehension for patients with limited health literacy or digital skills. One participant shared:</p><disp-quote><p>My parents, when they read their medical history and the doctor&#x2019;s notes, have no clue what any of it means. They have to put it in ChatGPT so it could be easier to understand.</p></disp-quote><p>However, participants emphasized that the effectiveness of AI in these roles depends on its accessibility, adaptability to diverse patient populations, and ability to integrate seamlessly into existing systems. Barriers, such as digital literacy gaps and language differences, were flagged as critical considerations for designing inclusive AI tools.</p></sec><sec id="s3-4-5"><title>Theme 5: Privacy&#x2014;Patient Concerns Around Data Protection, Privacy, and Security</title><p>Concerns about data privacy and security were prominent across all scenarios, with participants expressing apprehension about how their sensitive health information would be stored, accessed, and used. Participants in the digital scribe scenario were particularly concerned about data transfer and storage, with one asking:</p><disp-quote><p>Is it transferable? Is it something that would stay within my healthcare unit, or could others access it?</p></disp-quote><p>Another participant queried:</p><disp-quote><p>How is it stored? Is it going to be posted on the patient portal? Are we able to access it?</p></disp-quote><p>Participants also raised concerns about potential misuse of data, particularly for non&#x2013;care-related purposes. In the virtual human scenario, 1 participant worried about the possibility of data mining, asking:</p><disp-quote><p>Would it be a gain for me, like if it was mining my data?</p></disp-quote><p>Similarly, in the portal messaging scenario, a participant asked:</p><disp-quote><p>How much does the chatbot know?</p></disp-quote><p>These concerns underscore a broader mistrust of data handling practices and the need for robust privacy protocols to safeguard patient information. Participants consistently called for systems to prioritize transparency and informed consent regarding data collection and usage, ensuring that personal information is used solely for its intended purposes.</p></sec></sec><sec id="s3-5"><title>Participant Evaluation of the Co-Design Sessions</title><p>Participants provided largely positive evaluations of the co-design session (<xref ref-type="table" rid="table4">Table 4</xref>).</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Participant evaluation of the artificial intelligence (AI) co-design workshop (N=17).</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Overall experience</td><td align="left" valign="bottom" colspan="5">Response, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top"/><td align="char" char="." valign="top">1</td><td align="char" char="." valign="top">2</td><td align="char" char="." valign="top">3</td><td align="char" char="." valign="top">4</td><td align="char" char="." valign="top">5</td></tr><tr><td align="left" valign="top">How would you rate the meeting overall?<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">2 (12)</td><td align="char" char="." valign="top">15 (88)</td></tr><tr><td align="left" valign="top">In general, how useful was the meeting?<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">4 (24)</td><td align="char" char="." valign="top">13 (77)</td></tr><tr><td align="left" valign="top">What did you think about the materials presented and discussed during the meeting?<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">5 (29)</td><td align="char" char="." valign="top">12 (71)</td></tr><tr><td align="left" valign="top">What did you think about the guideline/recommendation discussion?<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">1 (7)</td><td align="char" char="." valign="top">5 (33)</td><td align="char" char="." valign="top">9 (60)</td></tr><tr><td align="left" valign="top">How much did the meeting contribute to a shared awareness of AI and diagnostic safety?<sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">3 (6)</td><td align="char" char="." valign="top">8 (47)</td><td align="char" char="." valign="top">8 (47)</td></tr><tr><td align="left" valign="top">How much say did you feel you had in the discussion?<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">2 (12)</td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">15 (88)</td></tr><tr><td align="left" valign="top">Do you think that the opinions of the different stakeholders that were present at the meeting were all taken into consideration?<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="char" char="." valign="top">0 (0)</td><td align="char" char="." valign="top">1 (6)</td><td align="char" char="." valign="top">3 (18)</td><td align="char" char="." valign="top">6 (35)</td><td align="char" char="." valign="top">7 (41)</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>1=Poor, 5=Excellent.</p></fn><fn id="table4fn2"><p><sup>b</sup>1=Not useful, 5=Very useful.</p></fn><fn id="table4fn3"><p><sup>c</sup>1=Almost nothing, 5=A lot.</p></fn><fn id="table4fn4"><p><sup>d</sup>1=Not at all, 5=Very much.</p></fn></table-wrap-foot></table-wrap><p>Participants highlighted several aspects of the workshop that they liked best. They appreciated the open and nonjudgmental environment, which allowed for free sharing of opinions and thoughts without bias or pressure. Many valued the interactive nature of the session, particularly the small group discussions, which facilitated deeper engagement, diverse perspectives, and meaningful participation. The diverse backgrounds of participants, including patients from different races and professions, enriched the discussions and provided new insights. Participants also found the materials well-prepared, appreciated the brief AI introduction, and enjoyed the opportunity to learn more about AI in relation to their health care. Overall, the combination of open dialogue, group interaction, and thoughtful organization was highly praised.</p><p>Participants shared a few areas for improvement in the workshop. The most common concern was the limited time available, with several noting the need for more time to discuss topics in greater depth and brainstorm ideas. Some also suggested dedicating additional time for group discussions and addressing specific examples of AI currently in use or relevant case studies. Suggestions included incorporating more complex cases or scenarios, discussing AI bias in greater detail, and diversifying both the researcher backgrounds and participant groups to include more primary care providers, individuals from different socioeconomic groups, and a broader generational representation. While some recommended separating patients based on their AI knowledge for tailored discussions, others emphasized maintaining a mix of diverse perspectives within groups. Overall, participants highlighted opportunities to enhance inclusivity, depth of discussion, and time for meaningful engagement.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Results</title><p>AI, while previously a technology of the future, has become a technology of the present. AI-driven technologies, including machine learning-driven decision support algorithms, deep-learning radiology scan classifiers, and large language model&#x2013;driven digital scribes, have already been implemented in hundreds of hospitals nationwide. This study, in examining patient perceptions of 5 different scenarios describing current and future AI technologies in health care, provides a contemporary view of the multifaceted patient perspectives on AI&#x2019;s role in providing diagnostic information, facilitating communication, and supporting decision-making. Many patient perceptions held true across all scenarios. First, a need for transparency in the development and validation of AI models, including their ability to reliably address the diverse needs of patients. Second, a preference for AI to complement rather than replace human providers, with an emphasis on maintaining human oversight in clinical decision-making. Third, the importance of clear and respectful communication about AI&#x2019;s role in care, including obtaining patient consent, was seen as essential for building trust. Fourth, the potential for AI to enhance patient engagement, understanding, and access to information, provided it is implemented as a supportive tool that respects patient autonomy. Finally, concerns regarding the security and privacy of patient data highlight the need for transparency and robust safeguards to prevent misuse or unauthorized access.</p><p>The outlook on AI implementation into the diagnostic process was generally positive, with participants consistently highlighting AI&#x2019;s ability to identify patterns and provide supplemental diagnostic information that might otherwise be overlooked by human providers. However, they emphasized that AI tools must integrate seamlessly into clinical workflows and preserve the essential human connection in patient-provider communication, as seen in the results regarding patient comfort levels across scenarios. Participants reported lower average comfort levels with high AI-patient interaction scenarios. For example, the scenario participants were least comfortable with was the virtual human scenario in which an AI-generated human would replace the physician when communicating a new diagnosis. Similarly, participants also appeared less comfortable with an AI chatbot sharing details about laboratory results. In contrast, participants were most comfortable with the digital scribe scenario, in which the application of AI was intended to enhance patient-provider communication by removing the need for providers to focus on documentation during clinic visits.</p><p>These findings highlight the importance of understanding patient perspectives within their specific health care contexts, offering insights into how AI can be integrated to enhance diagnostic processes and communication. For example, rural communities, where health care access and infrastructure often differ from urban settings, may present unique opportunities for AI to address gaps in care. Designing AI systems that are adaptable to varying levels of digital literacy, resource availability, and cultural expectations can help ensure these tools are both effective and equitable across diverse populations. Similarly, the interactive focus group discussions demonstrated the value of engaging patients with varying experiences and levels of familiarity with health care technologies. This diversity of input underscores the potential for AI to be developed in ways that resonate with patients from different backgrounds, fostering trust and satisfaction. By actively seeking and incorporating a wide range of perspectives, AI tools can be tailored to address the specific needs of different communities, ultimately supporting a more inclusive and patient-centered approach to care.</p></sec><sec id="s4-2"><title>Comparison With Previous Work</title><p>The findings of our work that held true across all scenarios are expectedly consistent with previous work &#x2013; patients&#x2019; concerns with privacy, data security, and bias have been well-documented [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. Specifically, our findings align well with the findings that patients have a generally positive outlook on AI&#x2019;s implementation into their care as long as there are adequate guardrails to protect against a variety of potential harms [<xref ref-type="bibr" rid="ref28">28</xref>]. Our work was unique in its focus on how such concerns are explicitly perceived in the context of AI applied to diagnosis and its communication in a variety of clinical contexts. Few qualitative studies have explored patient perceptions of AI in the context of diagnosis and communication, but our results appear to be in alignment with previous findings. Patients considered AI to be a helpful supplementary tool that should not serve as a replacement to human clinicians, a sentiment already documented for applications of AI in radiology [<xref ref-type="bibr" rid="ref23">23</xref>].</p><p>While there is previous work identifying patient perceptions on the implementation of AI in health care broadly, there has been limited work identifying patient perceptions of AI&#x2019;s role in reducing diagnostic errors through the enhancement of patient-provider communication. A recent scoping review identified that patients&#x2019; attitudes toward AI (which may impact their experiences when they interact with these tools in practice) are influenced by various factors, including familiarity with function, previous exposure to similar tools, supervision during use, and tool simplicity, validity, and cost. In light of this, it is imperative to consider patient perceptions of AI applications in a variety of clinical workflows in the context of diagnostic communication [<xref ref-type="bibr" rid="ref17">17</xref>]. As AI becomes ubiquitous both inside and outside of health care, patients&#x2019; familiarity (and thus their attitudes toward AI) will continue to evolve. It is critical that patient perceptions of AI tools are continually assessed and used to enhance the diagnostic process and communication so that AI can be designed and integrated into the health care system in such a way that maximizes patient care and satisfaction.</p></sec><sec id="s4-3"><title>Implications at the Individual Clinician, Organizational, and Policy or Regulatory Levels</title><p>Our findings have clear implications for clinical practice, finding that patients consistently stressed the importance of clinicians playing a central role in facilitating their experience with AI tools. Patients emphasized that their trust in AI would be built through transparent communication and clinician endorsement. Patients valued clinician involvement in contextualizing AI&#x2019;s outputs, interpreting its recommendations, and providing assurances about its accuracy and reliability. Patients additionally expressed concerns that AI could disrupt workflows or reduce human interaction, particularly in scenarios where key diagnostic information was to be communicated. To address this, organizations should prioritize the implementation of AI tools that enhance human connection in care, such as those that reduce administrative burdens. These tools should be designed in ways that enhance, rather than overshadow, the clinician&#x2019;s role in communicating diagnoses.</p><p>From a policy perspective, our findings reinforce the urgency of addressing gaps in regulations governing AI in health care, particularly concerning equity [<xref ref-type="bibr" rid="ref29">29</xref>]. Patients expressed concern that biased data inputs could undermine the diagnostic accuracy of AI tools and have harmful effects on historically underserved populations. Policies should ensure that tools are trained on diverse datasets and are validated across representative populations in order to build patient trust and acceptance. Patients also expressed calls for transparency and informed consent regarding how health data is collected and used by AI systems in health care, aligning well with findings from a study [<xref ref-type="bibr" rid="ref30">30</xref>]. These concerns should be addressed by policies that establish clear standards for disclosing when AI is used in care, and for ensuring that patients understand what personal information is being used, how it contributes to the diagnostic process, and how it influences the outputs from AI systems. Strengthening privacy and communication protocols will not only help address these concerns but also reinforce patient autonomy and trust in AI by supporting its ethical implementation in clinical settings.</p></sec><sec id="s4-4"><title>Importance of Involving Patients in AI Deployment</title><p>Our findings underscore the critical role patients play in the acceptance and success of AI tools designed to enhance the diagnostic process, emphasizing the need to involve them in the development and implementation of these technologies. As primary stakeholders most directly impacted by changes to diagnostic workflows, patients have invaluable insights that can guide the design of tools to align with their expectations and foster trust [<xref ref-type="bibr" rid="ref31">31</xref>]. Participants in this study expressed a dynamic view of AI tool implementation, with key insights into concerns that should be addressed during design and implementation, such as the importance of human connection and interaction, concerns regarding equity, personalization, and data security, and the pivotal role clinicians have in their understanding and comfort with new technology. By involving patients in the development of these tools, the health care system can better anticipate risks, communicate with patients more effectively, and deploy tools that not only improve the diagnostic process but also enhance trust and adoption, ensuring alignment with patient values and priorities.</p></sec><sec id="s4-5"><title>Limitations</title><p>This study has several limitations. First, focus groups were guided by different facilitators, which may have introduced variability in discussions due to differences in facilitation styles. The breakout sessions included only 2-4 participants per group, which allowed for in-depth discussions but may have limited the diversity of viewpoints within each session. This could affect the robustness of the findings; however, insights were aggregated across groups to capture broader themes. Future studies could address this by increasing group sizes or incorporating complementary methods such as individual interviews. Time constraints limited the discussion duration for each scenario, potentially restricting exploration of nuanced perspectives and the ability to achieve thematic saturation. In addition, the use of specific diagnostic-related AI scenarios provides structure but limits the generalizability of findings to other clinical contexts.</p><p>Participants were recruited through advisory networks, which may have introduced selection bias favoring individuals with an interest or familiarity in health care technologies. The participant demographics also reflect limitations in diversity, as more than 70% (n=13) were women, almost 65% (n=11) held graduate degrees, and none identified as having a Hispanic background. In addition, no participants reported high school or lower as their highest education level. According to national data, approximately 62% (n=115,011) of individuals aged 25 years and older in the United States have not attained a bachelor&#x2019;s degree, suggesting that our sample overrepresented highly educated individuals [<xref ref-type="bibr" rid="ref32">32</xref>]. Perspectives from participants with less formal education or from underrepresented backgrounds may have differed significantly, potentially revealing lower trust in AI, different concerns about its use, or alternate expectations for its role in health care. This lack of demographic diversity may limit the generalizability of the findings, as perspectives from individuals with different educational or cultural backgrounds could provide unique insights into AI applications in health care. We also did not assess baseline AI familiarity or digital literacy, which may have influenced participant engagement. Finally, this study is not comprehensive of all patient concerns about AI, with a focus on diagnostic applications shaping the discussions. Future research should aim to include a more demographically representative sample, explicitly assess AI literacy, and explore a wider range of clinical and nonclinical AI use cases to better understand how diverse patient populations perceive and respond to its implementation in the health care setting.</p></sec><sec id="s4-6"><title>Conclusions</title><p>This study highlights the nuanced perspectives of patients on the use of AI in health care, with a particular focus on diagnostic communication. While participants recognized the potential of AI to improve diagnostic accuracy, efficiency, and equity, they also voiced significant concerns about transparency, trust, and the preservation of human connection. These findings underscore the importance of ensuring that AI tools are developed and integrated in ways that align with patient values and priorities. Key patient-oriented considerations include the need for clear communication about AI&#x2019;s role in care, consent processes for its use, and opportunities for patients to actively participate in its development and implementation. Participants emphasized the importance of maintaining provider oversight, fostering understanding through accessible explanations, and designing systems that prioritize inclusivity and respect for patient autonomy.</p><p>As AI technologies continue to evolve and permeate health care, it is essential to iteratively assess and incorporate patient feedback to ensure these tools not only meet technical and clinical standards but also uphold the values of equity, transparency, and shared decision-making. By centering patients in the design and deployment of AI, we can create systems that not only enhance health care delivery but also foster trust and meaningful engagement between patients and providers.</p></sec></sec></body><back><ack><p>This project was funded under a Diagnostic Safety Centers of Excellence grant (1R18HS029356) from the Agency for Healthcare Research and Quality (AHRQ), US Department of Health and Human Services (HHS). The authors are solely responsible for this document&#x2019;s contents, findings, and conclusions, which do not necessarily represent the views of AHRQ. Readers should not interpret any statement in this report as an official position of AHRQ or HHS. None of the authors has any affiliation or financial involvement that conflicts with the material presented in this report.</p></ack><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hajkowicz</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sanderson</surname><given-names>C</given-names> </name><name name-style="western"><surname>Karimi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bratanova</surname><given-names>A</given-names> </name><name name-style="western"><surname>Naughtin</surname><given-names>C</given-names> </name></person-group><article-title>Artificial intelligence adoption in the physical sciences, natural sciences, life sciences, social sciences and the arts and humanities: a bibliometric analysis of research publications from 1960-2021</article-title><source>Technol Soc</source><year>2023</year><month>08</month><volume>74</volume><fpage>102260</fpage><pub-id pub-id-type="doi">10.1016/j.techsoc.2023.102260</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Du-Harpur</surname><given-names>X</given-names> </name><name name-style="western"><surname>Watt</surname><given-names>FM</given-names> </name><name name-style="western"><surname>Luscombe</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Lynch</surname><given-names>MD</given-names> </name></person-group><article-title>What is AI? Applications of artificial intelligence to dermatology</article-title><source>Br J Dermatol</source><year>2020</year><month>09</month><volume>183</volume><issue>3</issue><fpage>423</fpage><lpage>430</lpage><pub-id pub-id-type="doi">10.1111/bjd.18880</pub-id><pub-id pub-id-type="medline">31960407</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sunarti</surname><given-names>S</given-names> </name><name name-style="western"><surname>Fadzlul Rahman</surname><given-names>F</given-names> </name><name name-style="western"><surname>Naufal</surname><given-names>M</given-names> </name><name name-style="western"><surname>Risky</surname><given-names>M</given-names> </name><name name-style="western"><surname>Febriyanto</surname><given-names>K</given-names> </name><name name-style="western"><surname>Masnina</surname><given-names>R</given-names> </name></person-group><article-title>Artificial intelligence in healthcare: opportunities and risk for future</article-title><source>Gac Sanit</source><year>2021</year><volume>35 Suppl 1</volume><fpage>S67</fpage><lpage>S70</lpage><pub-id pub-id-type="doi">10.1016/j.gaceta.2020.12.019</pub-id><pub-id pub-id-type="medline">33832631</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hardy</surname><given-names>M</given-names> </name><name name-style="western"><surname>Harvey</surname><given-names>H</given-names> </name></person-group><article-title>Artificial intelligence in diagnostic imaging: impact on the radiography profession</article-title><source>Br J Radiol</source><year>2020</year><month>04</month><day>1</day><volume>93</volume><issue>1108</issue><fpage>31821024</fpage><pub-id pub-id-type="doi">10.1259/bjr.20190840</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Artificial intelligence in clinical decision support systems for oncology</article-title><source>Int J Med Sci</source><year>2023</year><volume>20</volume><issue>1</issue><fpage>79</fpage><lpage>86</lpage><pub-id pub-id-type="doi">10.7150/ijms.77205</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shaik</surname><given-names>T</given-names> </name><name name-style="western"><surname>Tao</surname><given-names>X</given-names> </name><name name-style="western"><surname>Higgins</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Remote patient monitoring using artificial intelligence: current state, applications, and challenges</article-title><source>WIREs Data Min &#x0026; Knowl</source><year>2023</year><month>03</month><volume>13</volume><issue>2</issue><fpage>e1485</fpage><pub-id pub-id-type="doi">10.1002/widm.1485</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Juhn</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name></person-group><article-title>Artificial intelligence approaches using natural language processing to advance EHR-based clinical research</article-title><source>J Allergy Clin Immunol</source><year>2020</year><month>02</month><volume>145</volume><issue>2</issue><fpage>463</fpage><lpage>469</lpage><pub-id pub-id-type="doi">10.1016/j.jaci.2019.12.897</pub-id><pub-id pub-id-type="medline">31883846</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Auerbach</surname><given-names>AD</given-names> </name></person-group><article-title>Laying a foundation for the use of artificial intelligence in diagnosis</article-title><source>JAMA Netw Open</source><year>2024</year><month>09</month><day>3</day><volume>7</volume><issue>9</issue><fpage>e2431907</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2024.31907</pub-id><pub-id pub-id-type="medline">39250158</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Newman-Toker</surname><given-names>DE</given-names> </name><name name-style="western"><surname>Nassery</surname><given-names>N</given-names> </name><name name-style="western"><surname>Schaffer</surname><given-names>AC</given-names> </name><etal/></person-group><article-title>Burden of serious harms from diagnostic error in the USA</article-title><source>BMJ Qual Saf</source><year>2024</year><month>01</month><day>19</day><volume>33</volume><issue>2</issue><fpage>109</fpage><lpage>120</lpage><pub-id pub-id-type="doi">10.1136/bmjqs-2021-014130</pub-id><pub-id pub-id-type="medline">37460118</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pushadapu</surname><given-names>N</given-names> </name></person-group><article-title>AI-enhanced techniques for pattern recognition in radiology imaging: applications, models, and case studies</article-title><source>JBAI</source><year>2022</year><month>03</month><volume>2</volume><issue>1</issue><fpage>153</fpage><lpage>190</lpage></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Reddy</surname><given-names>A</given-names> </name><name name-style="western"><surname>Reddy</surname><given-names>RP</given-names> </name><name name-style="western"><surname>Roghani</surname><given-names>AK</given-names> </name><etal/></person-group><article-title>Artificial intelligence in Parkinson&#x2019;s disease: early detection and diagnostic advancements</article-title><source>Ageing Res Rev</source><year>2024</year><month>08</month><volume>99</volume><issue>102410</issue><fpage>102410</fpage><pub-id pub-id-type="doi">10.1016/j.arr.2024.102410</pub-id><pub-id pub-id-type="medline">38972602</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Giordano</surname><given-names>C</given-names> </name><name name-style="western"><surname>Brennan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Mohamed</surname><given-names>B</given-names> </name><name name-style="western"><surname>Rashidi</surname><given-names>P</given-names> </name><name name-style="western"><surname>Modave</surname><given-names>F</given-names> </name><name name-style="western"><surname>Tighe</surname><given-names>P</given-names> </name></person-group><article-title>Accessing artificial intelligence for clinical decision-making</article-title><source>Front Digit Health</source><year>2021</year><volume>3</volume><issue>645232</issue><fpage>645232</fpage><pub-id pub-id-type="doi">10.3389/fdgth.2021.645232</pub-id><pub-id pub-id-type="medline">34713115</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cohen</surname><given-names>M</given-names> </name><name name-style="western"><surname>Puntonet</surname><given-names>J</given-names> </name><name name-style="western"><surname>Sanchez</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Artificial intelligence vs. radiologist: accuracy of wrist fracture detection on radiographs</article-title><source>Eur Radiol</source><year>2023</year><month>06</month><volume>33</volume><issue>6</issue><fpage>3974</fpage><lpage>3983</lpage><pub-id pub-id-type="doi">10.1007/s00330-022-09349-3</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zheng</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Zeng</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Artificial intelligence performance in detecting tumor metastasis from medical radiology imaging: a systematic review and meta-analysis</article-title><source>EClinicalMedicine</source><year>2021</year><month>01</month><volume>31</volume><issue>100669</issue><fpage>100669</fpage><pub-id pub-id-type="doi">10.1016/j.eclinm.2020.100669</pub-id><pub-id pub-id-type="medline">33392486</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ouanes</surname><given-names>K</given-names> </name><name name-style="western"><surname>Farhah</surname><given-names>N</given-names> </name></person-group><article-title>Effectiveness of artificial intelligence (AI) in clinical decision support systems and care delivery</article-title><source>J Med Syst</source><year>2024</year><month>08</month><day>12</day><volume>48</volume><issue>1</issue><fpage>39133332</fpage><pub-id pub-id-type="doi">10.1007/s10916-024-02098-4</pub-id><pub-id pub-id-type="medline">39133332</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hoseini</surname><given-names>M</given-names> </name></person-group><article-title>Patient experiences with AI in healthcare settings</article-title><source>aitechbesosci</source><year>2023</year><month>07</month><volume>1</volume><issue>3</issue><fpage>12</fpage><lpage>18</lpage><pub-id pub-id-type="doi">10.61838/kman.aitech.1.3.3</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Moy</surname><given-names>S</given-names> </name><name name-style="western"><surname>Irannejad</surname><given-names>M</given-names> </name><name name-style="western"><surname>Manning</surname><given-names>SJ</given-names> </name><etal/></person-group><article-title>Patient perspectives on the use of artificial intelligence in health care: a scoping review</article-title><source>J Patient Cent Res Rev</source><year>2024</year><volume>11</volume><issue>1</issue><fpage>51</fpage><lpage>62</lpage><pub-id pub-id-type="doi">10.17294/2330-0698.2029</pub-id><pub-id pub-id-type="medline">38596349</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Haggenm&#x00FC;ller</surname><given-names>S</given-names> </name><name name-style="western"><surname>Krieghoff-Henning</surname><given-names>E</given-names> </name><name name-style="western"><surname>Jutzi</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Digital natives&#x2019; preferences on mobile artificial intelligence apps for skin cancer diagnostics: survey study</article-title><source>JMIR Mhealth Uhealth</source><year>2021</year><month>08</month><day>27</day><volume>9</volume><issue>8</issue><fpage>e22909</fpage><pub-id pub-id-type="doi">10.2196/22909</pub-id><pub-id pub-id-type="medline">34448722</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jutzi</surname><given-names>TB</given-names> </name><name name-style="western"><surname>Krieghoff-Henning</surname><given-names>EI</given-names> </name><name name-style="western"><surname>Holland-Letz</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Artificial intelligence in skin cancer diagnostics: the patients&#x2019; perspective</article-title><source>Front Med (Lausanne)</source><year>2020</year><volume>7</volume><issue>233</issue><fpage>233</fpage><pub-id pub-id-type="doi">10.3389/fmed.2020.00233</pub-id><pub-id pub-id-type="medline">32671078</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lennox-Chhugani</surname><given-names>N</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Pearson</surname><given-names>V</given-names> </name><name name-style="western"><surname>Trzcinski</surname><given-names>B</given-names> </name><name name-style="western"><surname>James</surname><given-names>J</given-names> </name></person-group><article-title>Women&#x2019;s attitudes to the use of AI image readers: a case study from a national breast screening programme</article-title><source>BMJ Health Care Inform</source><year>2021</year><month>03</month><volume>28</volume><issue>1</issue><fpage>e100293</fpage><pub-id pub-id-type="doi">10.1136/bmjhci-2020-100293</pub-id><pub-id pub-id-type="medline">33795236</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ongena</surname><given-names>YP</given-names> </name><name name-style="western"><surname>Yakar</surname><given-names>D</given-names> </name><name name-style="western"><surname>Haan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kwee</surname><given-names>TC</given-names> </name></person-group><article-title>Artificial intelligence in screening mammography: a population survey of women&#x2019;s preferences</article-title><source>J Am Coll Radiol</source><year>2021</year><month>01</month><volume>18</volume><issue>1 Pt A</issue><fpage>79</fpage><lpage>86</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2020.09.042</pub-id><pub-id pub-id-type="medline">33058789</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kosan</surname><given-names>E</given-names> </name><name name-style="western"><surname>Krois</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wingenfeld</surname><given-names>K</given-names> </name><name name-style="western"><surname>Deuter</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Gaudin</surname><given-names>R</given-names> </name><name name-style="western"><surname>Schwendicke</surname><given-names>F</given-names> </name></person-group><article-title>Patients&#x2019; perspectives on artificial intelligence in dentistry: a controlled study</article-title><source>J Clin Med</source><year>2022</year><month>04</month><day>12</day><volume>11</volume><issue>8</issue><fpage>2143</fpage><pub-id pub-id-type="doi">10.3390/jcm11082143</pub-id><pub-id pub-id-type="medline">35456236</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Citardi</surname><given-names>D</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Genc</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Shan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>X</given-names> </name></person-group><article-title>Patients&#x2019; perceptions of using artificial intelligence (AI)-based technology to comprehend radiology imaging data</article-title><source>Health Informatics J</source><year>2021</year><volume>27</volume><issue>2</issue><fpage>33913359</fpage><pub-id pub-id-type="doi">10.1177/14604582211011215</pub-id><pub-id pub-id-type="medline">33913359</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Xiao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>K</given-names> </name><name name-style="western"><surname>Werner</surname><given-names>N</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>K</given-names> </name><name name-style="western"><surname>Hendrix</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hemmelgarn</surname><given-names>C</given-names> </name></person-group><article-title>Co-design with patients for improving patient safety: strategies, barriers and pitfalls</article-title><source>Proc Hum Factors Ergon Soc Annu Meet</source><year>2023</year><month>09</month><volume>67</volume><issue>1</issue><fpage>633</fpage><lpage>638</lpage><pub-id pub-id-type="doi">10.1177/21695067231192416</pub-id><pub-id pub-id-type="medline">38213999</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="report"><person-group person-group-type="author"><name name-style="western"><surname>HM</surname><given-names>HH</given-names> </name><name name-style="western"><surname>Hemmelgarn</surname><given-names>C</given-names> </name><name name-style="western"><surname>Coffee</surname><given-names>S</given-names> </name><etal/></person-group><article-title>The patient&#x2019;s role in diagnostic safety and excellence: from passive reception towards co-design</article-title><year>2024</year><publisher-name>Agency for Healthcare Research and Quality</publisher-name></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Williamson</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Prybutok</surname><given-names>V</given-names> </name></person-group><article-title>Balancing privacy and progress: a review of privacy challenges, systemic oversight, and patient perceptions in AI-driven healthcare</article-title><source>Appl Sci (Basel)</source><year>2024</year><volume>14</volume><issue>2</issue><fpage>675</fpage><pub-id pub-id-type="doi">10.3390/app14020675</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Beets</surname><given-names>B</given-names> </name><name name-style="western"><surname>Newman</surname><given-names>TP</given-names> </name><name name-style="western"><surname>Howell</surname><given-names>EL</given-names> </name><name name-style="western"><surname>Bao</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>S</given-names> </name></person-group><article-title>Surveying public perceptions of artificial intelligence in health care in the United States: systematic review</article-title><source>J Med Internet Res</source><year>2023</year><month>04</month><day>4</day><volume>25</volume><fpage>e40337</fpage><pub-id pub-id-type="doi">10.2196/40337</pub-id><pub-id pub-id-type="medline">37014676</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Richardson</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>C</given-names> </name><name name-style="western"><surname>Curtis</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Patient apprehensions about the use of artificial intelligence in healthcare</article-title><source>NPJ Digit Med</source><year>2021</year><month>09</month><day>21</day><volume>4</volume><issue>1</issue><fpage>34548621</fpage><pub-id pub-id-type="doi">10.1038/s41746-021-00509-1</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ratwani</surname><given-names>RM</given-names> </name><name name-style="western"><surname>Sutton</surname><given-names>K</given-names> </name><name name-style="western"><surname>Galarraga</surname><given-names>JE</given-names> </name></person-group><article-title>Addressing AI algorithmic bias in health care</article-title><source>JAMA</source><year>2024</year><month>10</month><day>1</day><volume>332</volume><issue>13</issue><fpage>1051</fpage><lpage>1052</lpage><pub-id pub-id-type="doi">10.1001/jama.2024.13486</pub-id><pub-id pub-id-type="medline">39230911</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>HJ</given-names> </name></person-group><article-title>Patient perspectives on informed consent for medical AI: a web-based experiment</article-title><source>Digit Health</source><year>2024</year><volume>10</volume><issue>20552076241247938</issue><fpage>20552076241247938</fpage><pub-id pub-id-type="doi">10.1177/20552076241247938</pub-id><pub-id pub-id-type="medline">38698829</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="report"><person-group person-group-type="author"><name name-style="western"><surname>Miller</surname><given-names>K</given-names> </name><name name-style="western"><surname>Biro</surname><given-names>J</given-names> </name><name name-style="western"><surname>Gold</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Hose</surname><given-names>BZ</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>H</given-names> </name><name name-style="western"><surname>Ratwani</surname><given-names>R</given-names> </name></person-group><article-title>Documenting diagnosis: exploring the impact of electronic health records on diagnostic safety</article-title><year>2024</year><access-date>2025-07-21</access-date><publisher-name>Agency for Healthcare Research and Quality</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.ahrq.gov/sites/default/files/wysiwyg/diagnostic/resources/issue-briefs/dxsafety-ehr-impact.pdf">https://www.ahrq.gov/sites/default/files/wysiwyg/diagnostic/resources/issue-briefs/dxsafety-ehr-impact.pdf</ext-link></comment></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="web"><article-title>Educational attainment in the United States: 2022</article-title><source>United States Census Bureau</source><year>2023</year><access-date>2024-08-22</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.census.gov/data/tables/2022/demo/educational-attainment/cps-detailed-tables.html">https://www.census.gov/data/tables/2022/demo/educational-attainment/cps-detailed-tables.html</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Artificial intelligence (AI) scenarios for breakout discussion.</p><media xlink:href="jopm_v17i1e69564_app1.docx" xlink:title="DOCX File, 15 KB"/></supplementary-material></app-group></back></article>