← All posts
Thoughts March 3, 2026

Stop Optimizing the Model. Start Securing the Runtime.

bdougie
bdougie Co-Founder

What model are you using? How does it benchmark? What’s your inference cost?

I get why these questions dominate every AI conversation right now. Models are visible, tangible, easy to compare. But optimizing for model performance while ignoring execution infrastructure is like choosing the fastest race car without checking whether the track has guardrails.

The real engineering challenge isn’t squeezing more capability out of LLMs. It’s figuring out how to actually run them safely. When an agent can write and execute code, you’re not dealing with AI-powered software anymore. You’re dealing with autonomous software running inside your infrastructure. Most stacks were not built for that.

Agents Execute Code. Your Stack Wasn’t Built for That.

Look at what engineering teams are actually building right now: micro-VM isolation, hardened sandboxes, credential proxying, disposable runtimes, strict environment stripping, backend separation between APIs and agents.

They’re building all of this because agents can access environment variables. Exfiltrate credentials. Modify infrastructure. Create runaway cost loops. Block production services.

“The real risk of autonomous AI isn’t hallucination. It’s execution. And that’s not a model problem. That’s an operating system problem.”

Isolation Is Converging. Observability Isn’t.

Sandboxing is converging fast. Run the agent in a container, strip secrets, proxy credentials, kill it if it misbehaves. Good work. But containment alone isn’t enough, and it’s not differentiated.

It’s also not sufficient. Most agent sandboxing today uses containers or microvms—Firecracker, Cloud Hypervisor, WASM runtimes. These work for disposable, stateless workloads. But they break down when you need secure boot, FIPS compliance, GPU passthrough for local model inference, or the ability for agents to run nested infrastructure like k8s or docker compose inside their own environment.

Microvms strip virtual hardware to optimize for boot speed. That tradeoff costs you secure boot, GPU access, bare metal support, and clean security boundaries for nested virtualization. For a demo, that’s fine. For a bank or a DOD contractor running agents in production, it’s a non-starter.

This is why stereOS uses full virtual machines. Each agent gets its own kernel, RAM, disk, and network stack. Nothing is shared with the host. The VM runs on real hardware, not just KVM, which matters when your enterprise customers need to self-host on bare metal.

The tradeoff people expect is boot speed. But stereOS optimizes for time-to-work, not time-to-interactive—a warm VM gets an agent to its first meaningful action in 70ms, without sacrificing any of the isolation guarantees that microvms throw away to get fast boot times.

But even with proper isolation, the harder question remains: if something goes wrong, can you actually prove what happened?

Can you replay the exact execution path? Reconstruct the full decision chain? Show what the agent read, wrote, and called? Provide a verifiable record of every state transition?

Most systems can’t. They isolate agents. They don’t make them inspectable. And opaque autonomy doesn’t scale. Not with enterprises. Not with auditors. Not with serious money behind it.

Even full session replay only gets you halfway. If humans still have to babysit agents, manually restart failed runs, debug unpredictable environments, and recover broken state, you don’t have autonomous systems. You have assisted automation with extra steps.

The real shift happens when agents can recover from failure, manage their own lifecycle, scale independently, and operate unattended—with every action remaining deterministic and verifiable. That’s not a logging problem. That’s a new compute model.

Agents Are a New Workload Class.

Right now, most teams are bolting agents onto existing infrastructure. An API here, a sandbox there, a proxy in the middle. It’s practical. It’s understandable. It’s also what things looked like before container runtimes existed.

New workload classes don’t get duct-taped onto old substrates forever. VMs changed what we could do with hardware. Containers changed how we shipped software. Serverless changed the unit of deployment. Each time, the durable companies were the ones that defined the new substrate layer.

Models are the hardware moment. Agent infrastructure is the substrate moment. The next category-defining companies won’t sit above the model. They’ll sit beneath it.

┌──────────────────────────────┐
│        Applications          │
├──────────────────────────────┤
│          Agents              │
├──────────────────────────────┤
│          Models              │  ← Commoditizing
├──────────────────────────────┤
│       Containment            │  ← Converging
│  (VMs, Docker, WASM)         │
├──────────────────────────────┤
│  Verifiable Execution        │
│  Deterministic Substrate     │
├──────────────────────────────┤
│        stereOS               │  ← Foundational Layer
└──────────────────────────────┘

What Happens When Agents Deploy Agents

A model gives you intelligence. It doesn’t give you deterministic execution, isolation guarantees, reproducible builds, capability scoping, verifiable state, or governance boundaries.

stereOS asks a different question than most of the tooling being built right now: what would an operating system look like if it were designed for autonomous agents from the beginning? Not a wrapper. Not middleware. Not another control plane bolted onto an existing stack. A substrate.

Here’s how it works. stereOS runs AI coding agents inside sandboxed Linux VMs. Each agent gets its own isolated kernel, memory, and disk. When you boot an agent, nothing is shared with the host. When you’re done, you tear it down.

The CLI is called mb. And the thing that makes it interesting isn’t what it does for you. It’s that agents can use it too.

mb doesn’t care whether a human or an agent is calling it. That’s a design choice, not an accident. Give an agent access to mb and it can spawn child agents, monitor the fleet, recover from failures, and report results—all inside isolated VMs, all with locked-down egress and readonly source mounts so nothing can interfere with anything else.

What makes this safe isn’t orchestration alone. It’s that every agent runs in a constrained, reproducible environment with clear execution boundaries. stereOS environments are built deterministically, so the exact runtime that produced a result can be recreated bit-for-bit.

Here’s a concrete example. You give one orchestrator agent a large task—refactoring a monorepo into microservices. The orchestrator analyzes the codebase, breaks the work down, writes child configs, and spawns specialized agents for each service. One agent handles auth. Another handles the API layer. Another handles database migrations. Each one runs in its own VM with its own resource budget.

When one child fails, the orchestrator reads the error log, diagnoses the root cause, destroys the failed instance, and redeploys with a fixed config.

“You went to bed. The fleet kept working.”

What makes this possible isn’t a clever orchestration SDK or a DAG compiler. It’s a Linux CLI. Agents are already good at CLIs. The interface is the same whether a human or an agent is at the keyboard. That’s the architecture.

The Substrate Layer

Models are not the moat. Containment is not the moat.

The moat is the substrate that makes autonomy safe enough to deploy and predictable enough to scale. Every computing era produces a layer that feels obvious in hindsight. VMs. Containers. Serverless. We’re at that moment for autonomous agents.

We need autonomous systems that are safe, reproducible, and governable. The companies that define this layer will sit beneath the models, not above them. That’s not middleware. That’s operating system territory.

stereOS is built for that layer. Try it today.

← Older Introducing stereOS: An operating system for your agent Newer → Agents Need Black Box Recorders

Join the waitlist

We are launching soon, subscribe for early access.