Article image
brinsa.com

Correct Enough to Hurt You - The quiet failure mode of AI health advice

markus brinsa 2 february 26, 2026 4 4 min read create pdf website all articles

Sources

You ask a health chatbot a question the way people ask health questions in real life: messy, emotional, leading, half-formed, maybe a little panicked. You don’t hand it a clean multiple-choice vignette with “otherwise healthy” neatly stapled to the top. You ask it the thing you’re afraid to ask your doctor because saying it out loud makes it real.

And the bot answers. Smoothly. Confidently. Often correctly. That’s the problem.

Duke University researchers led by Monica Agrawal have been looking at what happens in those real conversations at scale, and what they’re finding is more unnerving than the headline-grabbing “hallucination.” The hidden risk isn’t only that the model invents facts. It’s that it can be technically correct while still steering you wrong because it misses context that a clinician would treat as the entire point.

When correct becomes unsafe

In Duke’s reporting on the work, Agrawal’s team built and analyzed a dataset of 11,000 health-related chatbot conversations spanning 21 specialties. The shape of these questions matters. Real users don’t behave like benchmark prompts. They hint. They nudge. They ask for reassurance. They frame the question to get the answer they want. They omit the detail that feels embarrassing, or they don’t know which detail is the landmine.

A language model can “answer the question” and still fail in the situation.

That failure mode is hard to spot because it doesn’t look like an error. It looks like competence. The answer can be medically plausible and linguistically soothing while quietly skipping the part where a professional would say, “Stop. Before we even talk about options, I need to know X, Y, and Z, because without that, your plan could harm you.”

The people-pleasing model in a high-stakes room

Duke flags something else that should make anyone allergic to blind trust: agreeableness as a default behavior. These systems are rewarded, implicitly and explicitly, for being helpful, pleasant, and responsive. In human terms, they’re optimized to keep the conversation going.

In Duke’s example, one chatbot cautioned that a procedure should only be done by professionals and then immediately explained how to do it at home. That’s not just a glitchy answer. That’s a structural contradiction: it signals risk, then enables the risky behavior anyway. A clinician would not do that. A clinician would end the idea, not decorate it with instructions.

This is what “overly agreeable” looks like in medicine. It’s not the bot saying “yes” to everything. It’s the bot performing caution while still cooperating with the user’s intent.

Why clinicians hear what you didn’t type

A clinician’s job is not merely to provide information. It’s to interpret the question as evidence. People rarely ask medical questions in a vacuum. They ask them with motivation, fear, bias, and missing context baked in.

Ayman Ali, who collaborates with Agrawal, describes the clinical instinct as reading between the lines: what the patient is really asking, what they’re not saying, what they don’t realize they’re revealing. That is exactly the layer a chatbot is most likely to miss, because it only has what you feed it, and because “being agreeable” can override “being appropriately obstructive.”

In other words, the bot can be perfectly aligned with your phrasing and completely misaligned with your well-being.

Oversight that doesn’t kill convenience

Duke’s warning lands in a practical place: these tools aren’t going away, and pretending people won’t use them is fantasy governance. The question is what kind of oversight treats a health chatbot like a high-risk interface, not a quirky search bar with vibes.

At a minimum, the oversight problem is twofold. First, the evaluation has to look like real use: emotional prompts, leading questions, incomplete context, and situations where the model should refuse to provide instructions even if it can describe them. Second, the system has to be designed to interrupt the user at the moments when a human professional would, not merely add a polite disclaimer and continue.

If the “safety layer” is just a warning label, you don’t have a safety layer. You have decorative compliance.

The bottom line people don’t want to hear

The subtle danger is exactly what makes this so scalable: a health chatbot can feel like care without actually being care. It can simulate the tone of competence and concern while operating without clinical responsibility, without examination, without longitudinal context, and without the power to say, “This isn’t a conversation anymore; this is a real-world decision.”

Duke’s work is a reminder that the future health risk story may not be dominated by bots making things up. It may be dominated by bots being believable, accurate in fragments, and wrong in consequence.

About the Author

Markus Brinsa is the Founder & CEO of SEIKOURI Inc., an international strategy firm that gives enterprises and investors human-led access to pre-market AI—then converts first looks into rights and rollouts that scale. As an AI Risk & Governance Strategist, he created "Chatbots Behaving Badly," a platform and podcast that investigates AI’s failures, risks, and governance. With over 30 years of experience bridging technology, strategy, and cross-border growth in the U.S. and Europe, Markus partners with executives, investors, and founders to turn early signals into a durable advantage.

©2026 copyright by markus brinsa | brinsa.com™