Artificial Intelligence
Experts warn ai-generated health content risks misinterpretation without human oversight
Wednesday, April 15, 2026
|
Russ Scritchfield |
Why AI health guidance still needs human interpretation to avoid missteps. Nature Medicine Study Shows AI Health Advice Is Often Misunderstood; Experts Say Real People Still Crucial in Healthcare Communications
As AI generated personas and automated video hosts enter health communications, experts caution that scaling content without human interpretation can raise the odds that people misunderstand what they read or watch. Digital platforms have become a primary source for health information, with more than half of adults turning to social feeds for guidance and many consulting AI tools before speaking with a clinician. The volume and speed of AI make health knowledge more accessible, but access alone does not guarantee better decisions.
Nature Medicine Study Shows AI Health Advice Is Often Misunderstood; Experts Say Real People Still Crucial in Healthcare Communications
Recent findings underscore a clear gap between technical accuracy and human understanding. In controlled settings, AI systems could correctly identify medical conditions in roughly 95 percent of test cases. Yet when people used the same tools in realistic scenarios, they landed on the right answer less than 35 percent of the time and did not make better choices than those relying on non AI sources. The problem is not only whether the model is right. The bigger issue is how people share context, interpret outputs, and decide what to do next.
The communication gap matters more than model accuracy
The study linked errors to routine user behavior. Participants often provided incomplete details, misunderstood the response, or failed to follow through on correct suggestions. Even well written answers can go astray when a person leaves out a symptom, misreads risk, or assumes reassurance where caution was intended. Reported cases have also shown AI services missing the need for urgent escalation, which reinforces that guidance must be paired with clear next steps and easy handoffs to clinicians. This is a communication challenge as much as it is a modeling challenge.
Creators act as interpreters between brands and patients
For marketers and health brands, these findings are practical, not theoretical. Donatas Smailys, CEO of Billo App, notes that health content has unique stakes. He explains, Healthcare is one of the few categories online where being slightly wrong can have consequences, as brands are legally responsible for the information they put out. That is what sets the standard, but how that information is communicated still matters. In practice, consumers rarely encounter medical information in its raw form. Smailys adds, Creators often act as a layer between brands and audiences, shaping how information is presented. They question, simplify, and adapt messaging in a way that makes it more understandable. This is actually what separates good creators from others.
He cautions that new content pipelines may remove that interpretive layer. If brands start generating content directly through AI, and a script is simply fed to an AI avatar or influencer, that layer can disappear, Smailys says. Creators serve as interpreters, mediators for broader audiences, and without it, we would just be feeding some visual AI slop. That increases the risk that information is presented in a way people misunderstand. Experience also helps creators catch rough edges. As Smailys puts it, The fact that a human prompted it does not mean that it is going to be correct. An experienced healthcare creator would easily catch that something sounds too fake, slightly off, and quickly dive deeper to fix it.
AI personas and avatars raise new communication risks
Health leaders in the United States have floated the use of AI avatars to extend mental health services in communities that lack providers. Supporters argue this can expand access and offer timely responses. Smailys agrees that reliable information matters, but he stresses that accuracy alone is not enough. AI can generate accurate medical information, but most people do not have the background to know what to ask or how to interpret the answer, he says. Information is simply better conveyed when a human communicates it to another human. That human layer translates vocabulary, adds context, and checks whether the audience truly understands the next step.
Implications for marketers platforms and healthcare leaders
Better outcomes will come from how health information is delivered, not only how it is generated. Platforms and brands should not prioritize engagement at the cost of user health, Smailys says. Educational content can be just as engaging when it is done properly, and calling out misinformation should be part of that system. He argues that platforms need to rebalance incentives. If algorithms continue to reward reach without considering accuracy, the problem will scale. From Meta to TikTok, there needs to be a shift toward promoting content that is not just engaging, but also credible.
This shift includes supporting creators with medical input and elevating key opinion leaders who have clinical or scientific expertise. Programs that train creators alongside healthcare professionals have shown promise. In comparative evaluations, trained creators were more likely to include evidence based information than untrained peers, and a larger share of younger users reported that trained content was useful. The takeaway is straightforward. When creators have access to medical guidance and clear review processes, audience understanding improves.
Practical steps to reduce misunderstanding
Brands can take concrete actions now. Add a human review step that pairs medical expertise with audience savvy, so messages are both correct and clear. Test scripts with diverse users to catch confusing phrasing before publishing. Use prompts and interfaces that guide people to share complete, relevant information, and show limits and next steps in plain language. Build escalation pathways into experiences so that concerning symptoms or uncertainty lead to clinician contact. Measure success with comprehension checks and safe behavior outcomes, not clicks alone. Finally, be transparent about when AI is involved and make support easy to reach if something is unclear.
A balanced path forward
AI can help scale high quality health information, but people interpret and act on that information in social, messy, real world contexts. That is why creators, clinicians, and informed communicators remain essential. Pairing AI with human oversight respects both the strengths and limits of each. It also aligns incentives toward clarity, safety, and trust. As Smailys notes, Information is simply better conveyed when a human communicates it to another human. Health communication works best when technology provides reach and speed, and real people ensure that meaning carries through.
Become a subscriber of App Developer Magazine for just $5.99 a month and take advantage of all these perks.
MEMBERS GET ACCESS TO
- - Exclusive content from leaders in the industry
- - Q&A articles from industry leaders
- - Tips and tricks from the most successful developers weekly
- - Monthly issues, including all 90+ back-issues since 2012
- - Event discounts and early-bird signups
- - Gain insight from top achievers in the app store
- - Learn what tools to use, what SDK's to use, and more
Subscribe here
