Field guide · Healthcare AI · Updated April 2026

Healthcare AI agents,
explained by people who ship them.

A practitioner-grade guide to healthcare AI agents in 2026. The architectural choices that separate agents from chatbots, the deployment shapes that actually work in regulated environments, and the production proof points behind the claims. Written by the team running 30+ agents across HIP One and PES One — including the live CMS Medicare deployment under the WISeR Model.

See the marketplace Read the architecture
The short version

An AI agent is a specialist, not a chatbot.

A healthcare AI agent is a software agent — typically built on large language models grounded in clinical criteria — that handles a specific clinical or administrative job. It verifies eligibility (calls an API), evaluates criteria (runs a deterministic rule pack), assembles evidence (extracts structured data from clinical documents), produces an audit packet (logs every decision). The defining trait is decomposition — agents are specialists, not generalists. Production-grade systems run several specialist agents in sequence, not one big LLM.

Where AI agents work in healthcare

Six categories of production-grade healthcare AI agents.

Agents that are actually shipping in production today, in CMS-validated environments and at named health systems. Not the demo-stage stuff.

Category 1

Prior authorization agents

Eligibility, intake, duplicate detection, criteria evaluation, evidence assembly, auto-affirmation. The most mature category in 2026, with multiple production deployments at scale. See the agent guide.

Category 2

Utilization management agents

Concurrent review, retrospective review, length-of-stay prediction, level-of-care recommendations. The architectural cousin of PA agents — same evidence-grounding pattern, different decision points.

Category 3

Patient engagement agents

Voice (IVR replacement), digital chat, mobile, proactive outreach. Member identity, intent classification, knowledge retrieval, conversation orchestration. See the patient engagement guide.

Category 4

Clinical documentation agents

Ambient scribing, encounter note generation, ICD/CPT coding assistance, quality measure abstraction. The fastest-growing category, increasingly EMR-embedded.

Category 5

Privacy and compliance agents

Incident triage, BAA tracking, disclosure accounting analysis. Most CPS One workflows are deliberately deterministic — but pattern-detection across aggregated reporting data uses an Aether One™ analytical agent.

Category 6

Payment integrity agents

Claims edit recommendation, duplicate claim detection, coordination-of-benefits analysis. Adjacent to UM agents architecturally; different financial outcome.

The architectural decisions that matter

Five questions that separate production agents from demos.

Most healthcare AI vendor pitches don't survive these five questions. The questions are technical; the answers are architectural; the implications are regulatory.

Question 1

Is the agent grounded in deterministic rules, or just an LLM call?

An LLM call wrapped in a system prompt is not a healthcare-grade agent. Production agents combine LLM reasoning with deterministic rule packs that cite NCD/LCD/medical-policy text verbatim, with version control and audit traceability on every rule.

Question 2

What happens when the agent is wrong?

A wrong agent decision in healthcare is not a UX problem — it's a clinical, regulatory, and financial event. Production architectures route adverse decisions through human reviewers (auto-deny prohibited), emit audit packets with evidence chains, and support determination reversal workflows.

Question 3

Can the agent run on-premises or only in the cloud?

Federal health programs and high-security payer environments cannot send PHI to commercial cloud. Cloud-only agents cannot serve those buyers. Sovereign-deployable agents can.

Question 4

Does the agent handle FHIR, X12, HL7 natively?

Healthcare integration is unusually complex. An agent that requires custom integration for every payer or provider is not actually deployable at scale. Production agents speak the standards natively.

Question 5

Is the methodology defensible IP, or just an LLM wrapper?

If a healthcare AI agent's competitive moat is "we use OpenAI plus a system prompt," it's reproducible by any well-funded competitor in a quarter. Patent-protected agent architecture — methodology that holds up under USPTO examination — is what differentiates a defensible business from a feature. See the Aether One™ patent portfolio →

What buyers ask

Buyer's questions, answered.

The seven questions that surface in every healthcare AI agent evaluation.

What are healthcare AI agents?

Healthcare AI agents are software agents — typically built on large language models grounded in clinical criteria — that handle specific clinical or administrative jobs in healthcare workflows. Examples: an eligibility agent checks beneficiary enrollment; a criteria agent evaluates medical necessity; an evidence agent assembles clinical documentation; a member navigation agent answers patient questions. The defining trait is decomposition — agents are specialists, not generalists.

How are healthcare AI agents different from a healthcare chatbot?

A chatbot generates conversational text. An agent makes decisions, calls tools, takes actions. A healthcare AI agent can verify eligibility (call an API), evaluate criteria (run a deterministic rule pack), assemble evidence (extract structured data from clinical documents), and produce an audit packet — all in a single workflow. Most production healthcare AI today involves agents, not chatbots.

Are healthcare AI agents safe?

It depends entirely on the architecture. Agents that wrap a single LLM call are not safe for clinical decisions. Agents grounded in deterministic rule packs, with explicit evidence chains and human-in-the-loop on adverse decisions, are safe enough for production CMS Medicare deployment — which is exactly what Aether One™ does. The architectural pattern matters more than the model choice.

How many healthcare AI agents does a typical platform run?

HIP One's prior authorization workflow uses 7 agents in sequence: Intake, Eligibility, Duplicate Request, Criteria, Evidence, Auto-Affirmation, and Non-Affirm Research. PES One's patient engagement workflow uses additional agents for member identity, intent classification, knowledge retrieval, and conversation orchestration. The Aether One™ substrate as a whole orchestrates more than 30 distinct agents across HIP One and PES One.

Are healthcare AI agents HIPAA compliant?

Aether One™ agents are designed for HIPAA compliance: PHI never leaves the customer environment in sovereign deployments, model inference can run on-premises or in government cloud, audit logs are HIPAA-grade, and PHI is never used for model training. Compliance also requires the surrounding governance — which is what CPS One provides for AI systems handling PHI.

Can healthcare AI agents run on-premises?

Yes. Aether One™ Sovereign is the on-premises deployment of the Aether One™ substrate, including all agents. Knowledge Containment Architecture (KCA) keeps weights, knowledge, and decisions inside the customer perimeter — a requirement for federal health programs and high-security payer environments that cannot send PHI to commercial cloud.

Who builds healthcare AI agents?

Several vendors build healthcare AI agents at production scale: Genzeon Platforms (Aether One™ for PA, UM, patient engagement, privacy), Cohere Health (PA), Abridge (clinical documentation), Hippocratic AI (member-facing), Glass Health (clinical reasoning). The architectural patterns vary. Genzeon's distinguishing pattern is domain-decomposed agents, sovereign deployment, and patent-protected methodology — 10 patents filed across the substrate.

Ready to evaluate?

Test an agent on your workflow.

A focused, time-boxed pilot with the team running healthcare AI agents in CMS Medicare today. We bring the agents and the audit packets. You bring the use case.

Talk to the team Browse agents