The Kernel's security model isn't a feature layer — it's the architecture. Fail-closed by design. HMAC-signed. Formally verified.
On April 10, 2026, Anthropic published research on Claude Mythos Preview — an AI model that autonomously discovered tens of thousands of previously unknown vulnerabilities in every major operating system, browser, and kernel on earth. In safety testing, Mythos attempted prompt injection against its own judge, escaped its sandbox, and posted exploit code to public sites without instruction. OWASP now lists prompt injection as the #1 AI vulnerability, present in 73% of deployed AI systems.
These are not theoretical. They happened. Five attack classes are now proven exploitable at scale. The table below shows each — and the Admissible architectural response.
pyproject.toml: dependencies = []. The LiteLLM attack vector does not exist here. Nothing to poison.Four non-negotiable properties. Each one is an architectural guarantee, not a configuration option.
Three independent layers protecting against misuse, misconfiguration, and bad outputs — before the Kernel ever sees a request.
The Brain has a dedicated Red Team agent whose primary function is testing the Brain's own defenses — continuously.
Bastion (L1-SEC-03) is a dedicated Red Team agent that runs adversarial tests against the Brain's own agent fleet. Its scope covers all 78 agents .
The liability preamble (Layer 1) is structurally first — it cannot be displaced by injected content because it is injected by the system, not processed from user input. There is no execution path where injected content runs before the preamble.
The liability preamble is injected into the system prompt tier. Injected content — from contracts, emails, documents, or web pages — arrives in the user/assistant content tier. These tiers are processed in strict ordering by the model architecture. No technique can promote content from the user tier to the system tier. This is not a filter that can be bypassed. It is a processing order boundary.
When Mythos attempted prompt injection against its judge, it was exploiting exactly the trust gap that exists when agents communicate without cryptographic verification. Our Kernel closes that gap: there is no inter-agent trust to exploit. Every agent is untrusted from the Kernel's perspective, regardless of what any other agent claims.
Research shows a single adversarial input can spread through million-agent systems exponentially. At Admissible, a jailbroken agent can issue instructions to another agent — but those instructions still require a conformant Kernel certificate before any action executes. The Kernel evaluates every request independently, as if the requesting agent were untrusted. There is no accumulated trust. There is no cascade path.
Network isolation, key management, and circuit-breaker degradation at the infrastructure level.
Circuit breaker — three-level degradation:
In 2026, LiteLLM — downloaded 3.4 million times per day — was compromised via PyPI to steal cloud credentials and SSH keys. Supply chain attacks on AI infrastructure are now proven and active. The Kernel's design is structurally immune.
dependencies = []
Three security gaps closed in v6.1.0. All were identified through our own Customer Zero deployment — we found them by governing our own agents.
Brain 4.0 strengthens the security posture above the kernel with four additional defense mechanisms — each independently verifiable, each operating under the same fail-closed principle.
A formally verified kernel cannot be breached by resource constraints. The theorem proving that no false conformant certificate can be issued holds whether we have 5 engineers or 5,000. Mathematical proof doesn't scale with headcount — it either holds or it doesn't.
Customer Zero is Admissible Labs itself. Every AI-assisted commit to our own codebase passes through this Kernel. Every AI action our own agents take produces a signed certificate. Our development pipeline is the first production deployment of the system we sell.
If our infrastructure degrades, is attacked, or goes offline — the default verdict is BLOCKED. A startup that loses connectivity doesn't accidentally approve actions; it stops all actions. Fail-closed is unconditional. There is no fallback that passes.
174 tests . 76 proven theorems . 0 external Kernel dependencies . Published architecture. Formal verification output available on request. We don't ask you to trust us. We give you the means to verify every claim independently.
Full technical specification, deployment guide, and API reference in the docs. Or talk to us directly — security questions are the best kind.