The Braina chatbot is seen on a laptop. (Photo by Jaap Arriens/NurPhoto via Getty Images)
NurPhoto via Getty Images
According to a recent Gallup poll, sixteen percent of U.S. adults now use AI chatbots for medical advice. That’s a big shift in the patient journey — AI can help patients translate medical jargon and prepare for visits, but it also introduces real safety and privacy risks. Here are five practical guardrails patients can use to get the benefits without the harm.
Five Patient-Powered Guardrails For Safer AI Use
1) Understand Privacy Before You Paste Your Medical Record Into An AI Chatbot
Patients ask: “Should I load my chart into an AI chatbot for better answers?” Here’s what they need to think about:
Consumer AI chatbots (i.e., chatGPT, Gemini, Grok, Claude, etc) are not designed for medical privacy (HIPAA-compliant). If you share full records, you may be disclosing sensitive health information into systems governed by consumer terms—not healthcare privacy frameworks. Even Grok recently pushed back publicly at its creator, Elon Musk, warning it is not HIPAA compliant and not a substitute for professional medical advice, despite his calls to use AI for medical advice.
Your data may be used to improve models unless you take specific steps. OpenAI’s current guidance states that content from individual services (like ChatGPT) may be used to train models, and that users can opt out through a privacy portal or use “Temporary Chat” to avoid chat history/training use.
“Opting out” and “deleting” are not the same as “it never existed.” Policies vary by provider and can change quickly. For example, Anthropic’s recent privacy updates describe consumer settings that can allow chats to be used to improve Claude, and outline retention changes when users choose to allow training.
Practical patient rule: share the minimum information necessary (do not upload your entire record to an AI chatbot), remove identifiers (name, DOB, address, medical record number, full dates), and use privacy controls before pasting anything sensitive (opt-out/temporary chat). Assume screenshots and copy/paste can persist. Treat anything you enter as potentially retrievable later through logs, support processes, or policy changes. Lastly, if a tool itself warns it isn’t appropriate for medical privacy, take it seriously.
2) Force Reputable Sources—Or Require The Model To Refuse
AI offers a real opportunity for patients to translate complex jargon and prepare better questions for visits. However, a real risk exists: A recent University of Oxford study showcased in Forbes demonstrated that AI-based health advice is no better than online searches or personal judgment, and slight variations in symptom description can lead to contradictory urgency recommendations.
General consumer AI chatbots can blend high-quality medical references with low-quality content and even misinformation. Patients can reduce that risk by instructing the model to use only reputable sources and to cite them. Therefore, patients using consumer AI chatbots should minimize risk by forcing the AI to use and cite only reputable sources (e.g., CDC, NIH, Mayo Clinic). The AI should act like it’s performing “retrieval + grounding” and must state “I don’t know” if it cannot support a claim with a trusted source. Treat un-cited or unsupported answers as unreliable.
Example Prompt: “Interpret this result [lab/imaging study] and explain at an [X] grade reading level. Use only CDC, NIH, MedlinePlus, Mayo Clinic, Cleveland Clinic, WHO, or PubMed. Provide a clickable citation for every claim. State ‘I don’t know’ if you can’t find support. [Insert result text]“
3) Use AI For Translation And Preparation—Not Diagnosis And Treatment Decisions
AI tools can be very helpful for patients in navigating the complexities of medical information and communication, but they also carry significant risks if used for self-diagnosis or to replace professional medical advice. For example, Stanford researchers evaluating AI therapy chatbots found another version of the same risk: these systems can respond in ways that reinforce harmful patterns, including stigmatizing responses and failures to appropriately handle crisis-like scenarios (e.g., suicidal ideation prompts), where safe responses should redirect users toward help rather than enable harm.
Patients should understand the distinction between appropriate and high-risk uses of these tools.
Good uses: Patients can leverage AI to translate medical jargon, gain a better understanding of a condition, summarize a symptom timeline, draft messages for their doctor’s portal, and generate focused questions for a short visit.
High-risk uses: It is dangerous to use AI for self-diagnosis (e.g., “Do I have X?”), making medication or treatment changes (“Should I stop this medication?”), assessing the need for emergency and urgent care, or attempting to challenge or “litigate” a clinician’s judgment.
If an AI output suggests a change in medical action, the next step should be consulting a qualified clinician, pharmacist, or nurse line, not more prompting.
4) Install A Brake Pedal For Rabbit Holes
The most significant concern for clinicians isn’t just a single incorrect answer from AI, but rather the pattern of repetitive prompting that escalates patient anxiety—what is often called an AI “rabbit hole.”
Emergency physician Graham Walker, MD described a compelling case illustrating this challenge. A patient with a classic nerve compression was properly worked up and reassured in the ED. However, she returned the next day, convinced by ChatGPT that she was at risk of “kidney failure” and coached on what to say. Dr. Walker highlighted that the issue was the patient’s competing trust in the persistent, emotionally attuned, and always-available AI tool, which led to an “anxiety loop” and undermined professional medical advice.
This pattern is not limited to emergency care. In outpatient settings, I’ve seen a patient become fixated on a non-existent medication side effect after repeated chatbot prompting, which eroded reassurance and led to multiple unnecessary urgent-care visits.
These stories illustrate a crucial point: AI can help you prepare for care, but it should not become the care. To protect against these anxiety rabbit holes, patients can use a simple boundary:
Anxiety/Rabbit-Hole Guardrail
- If the chatbot increases your fear, urges you to ignore professional advice, or suggests repeated urgent visits without new symptoms, stop.
- If you’ve asked essentially the same question twice, and your anxiety is increasing, stop.
- If the chatbot is pushing repeat urgent visits without new symptoms, stop immediately and contact a clinician or nurse line.
5) Which AI Model Is Best For Medical Questions?
Most patients aren’t walking into clinic asking, “Should I use Claude or Gemini?” They’re using whatever is available and convenient—the free AI chatbot on their phone, the tool that pops up first in search, a patient portal AI chatbot or whatever a friend mentioned. And because both the foundation models and the healthcare features inside patient portals are changing so quickly, it’s hard for anyone—patients included—to keep up with what’s new, what’s safer, and what has real guardrails.
That’s why I think the most useful way to answer “which model should I use?” is to start with what kind of tool reduces risk by design. Health systems are beginning to embed AI chatbots inside patient portals that can be constrained to vetted content and your record—Epic, for example, has been rolling out a patient-facing AI called “Emmie.” Foundation model companies are also releasing health-specific experiences; OpenAI recently introduced ChatGPT Health. Several peer-reviewed studies compare LLMs on medical MCQs or specialty exams. Results vary by domain, language, and the specific model/version tested:
A cardiology MCQ study found Claude highest overall (78.3%), with ChatGPT-4 and Gemini close behind (75.9% each). A Scientific Reports study found performance differences across ChatGPT, Gemini and Claude vary by specialty and language.
Bottom Line
Don’t let model shopping become the project. Pick the option that’s most constrained and privacy-aware, then rely on guardrails.
AI can make medical information more accessible, but only with boundaries. Don’t upload full PHI, insist on reputable citations, and use AI to prepare — not to replace — clinician judgment. If an AI chatbot increases fear or contradicts your clinician, stop and bring the issue to your doctor.

Leave a comment