Challenge mode

Can your healthcare AI survive 10 traps?

Run a local-first healthcare AI prompt-injection challenge, get a survival verdict, inspect baseline agent context, and publish the result as a visual proof dossier, offline proof PDF, generated UI mockup, report, README marker, social card, SARIF file, OpenTelemetry log bundle, JSON, and Markdown.

Evidence loop

Honeypot Med has a repeatable proof mechanic.

A useful developer tool needs an outcome people can screenshot, review, and cite from a README.

Survival verdict

The default challenge pack returns a result like 8/10 survived, with blocked traps called out as concrete report events.

README marker

Every challenge bundle includes badge.svg and README-badge.md so other projects can link directly to their proof packet.

Baseline context

Reports include OpenAI-compatible chat, RAG bot, claims copilot, prior-auth agent, and voice-agent baseline profiles.

Visual packet

Every challenge also exports proof-dossier.html, offline-proof.pdf, and ui-mockup.html so the result is readable without a terminal.

Baselines

Compare against familiar agent shapes.

These are representative profiles, not vendor claims. They make the challenge legible to builders who know their architecture.