AI & Automation

Building HIPAA-compliant AI agents

An updated take on implementing no-code in healthcare, with the added realities of AI agents and examples of use cases in patient management.

Building HIPAA-compliant AI agents

Read this if you’re building HIPAA compliant software and systems and want to use AI agents

This is an update to a case study I originally shared on NoCode.tech about bringing no-code (now known as vibe-code) into a regulated healthcare org (patient portal + care-team workflows, built to move fast without letting PHI leak into the wrong tools). See the original write-up here: https://www.nocode.tech/article/i-helped-implement-no-code-with-a-healthcare-org-heres-what-i-learned

What’s different now is not HIPAA. It’s the shape of the risk when AI agents enter the system.

Context: the no-code healthcare case study this is based on

The baseline pattern is the same as the NoCode.tech article: keep PHI inside a HIPAA appropriate boundary, and still let a small care team move quickly with no code by pushing non-PHI work into simpler tools. The stack combined internal tools, automations, and patient communications. The rule was simple: if a tool cannot protect PHI, it should never see PHI.

What changed with AI agents

The temptation now is to add an agent to every workflow. In healthcare, that is only safe if you treat the agent like a new teammate.

Modern agent use cases (updated versions of the original workflows)

The original build used no-code tools to automate patient communication and give the care team leverage. With agents, the same workflows get smarter, but they also need tighter boundaries.

1) Patient outreach agent (SMS or email) for “didn’t complete X” follow-ups

Original idea: find patients who match a condition, then text them a reminder.

Agent upgrade:

2) Intake and forms agent embedded in a marketing site

Original idea: build the website anywhere, but embed a HIPAA-capable form tool for PHI.

Agent upgrade:

ai-generated-image.png

3) Care-team copilot for “one person supports hundreds of patients” operations

Original idea: a simple portal + internal tooling so a small team can handle volume.

Agent upgrade:

4) De-identified automation agent for non-PHI tools (the safe speed boost)

Original idea: anonymize data, then use non-HIPAA tools for convenience.

Agent upgrade:

ai-generated-image.png

5) Documentation and training agent (keeps you audit-ready)

Original idea: document the architecture and train the team.

Agent upgrade:

Below are the updated learnings, mapped to the original structure.

1. Audit your data flows, including agent actions

The old audit was: where does PHI exist, and where does it travel?

The updated audit is: where does PHI exist, where does it travel, and where could an agent accidentally send it?

Practical additions for 2026:

2. Permissions are the main battlefield

In the original article, I warned about row level permissions and not putting PHI in tools that cannot restrict access.

With agents, permissions matter even more because they can chain actions in seconds: read, decide, write, notify.

Updated best practices:

3. Observability needs to be PHI safe

Everyone wants better debugging for AI. The trap is logging too much.

Updated guidance:

4. Document and train, now with an agent playbook

Compliance is repeatable process, as demonstrated by the rise in “compliance-as-service” companies live Vanta. HIPAA is a set of processes and checkboxes for people and machines, now including agents. The concept is not new.

What is new:

A simple reference architecture for agentic workflows

If you want a usable mental model, start here:

ai-generated-image.png

FAQ: HIPAA and AI agents (quick answers)

Can I use AI agents with PHI under HIPAA?

Yes, but only if the full system is designed for HIPAA: the agent, its tool access, where data is stored, and every vendor that could receive PHI. Treat the agent like a teammate with credentials: least privilege, approvals for high impact actions, and audit trails.

Is ChatGPT HIPAA compliant?

The consumer app is not the default choice for PHI. In practice, HIPAA compliant use requires the right enterprise or API setup plus a signed BAA and a PHI safe architecture. If you cannot get a BAA for the exact services you use, do not send PHI.

What makes an AI vendor “HIPAA compliant”?

There is no magic badge. The practical checklist is: a signed BAA, clear data handling terms (retention, training, access), and controls you can enforce (encryption, access restrictions, logging, and deletion).

What is the biggest new risk when agents enter a healthcare workflow?

The biggest risk is uncontrolled blast radius: an agent that can see too much data or act in too many systems, plus accidental data egress (PHI leaking into prompts, logs, or third-party tools).

Permission sprawl is one common way this happens, but it’s not the only one. The practical goal is the same: keep PHI tightly bounded, keep tool access minimal, and make every action auditable.

What should we log for agentic workflows without leaking PHI?

Prefer structured audit logs over raw prompts: timestamps, event type, tool name, who requested the action, what record IDs were touched, approval outcomes, and redacted summaries. If you keep transcripts, store them in a controlled system with strict access and retention.

What is a safe starting architecture for an AI agent in healthcare?

Keep PHI in a HIPAA boundary, keep non PHI in simpler systems, and put a gateway in between that enforces allowlists, input schemas, and approvals. Add an immutable audit log that captures decisions and tool calls.

Ready to build AI workflows for your team?

Schedule a call