This page describes direction, not a timeline. There are no quarters, no “coming Q4”, no projected launch dates. The phases below are ordered by what each one makes possible, not when we promise to ship it. We'd rather move slower and tell you the truth than commit to a calendar and quietly miss it.
These are principles, not features. They constrain every future architectural decision. If we ship something that violates one of these, it's a regression and we owe you an explanation.
We hold your agent's memory, conversations, and operational state on your behalf. We do not analyse it, sell it, train on it, or ship features that require us to read it in plaintext.
Today inference runs on shared infrastructure (Fireworks under Zero Data Retention). The architecture is built so this dependency shrinks over time — first per-tenant, then on-device. We don't want to be a permanent middleman between you and a model.
“We promise not to look” is a procedural assurance — it depends on us being trustworthy and our access controls being correct. Cryptographic privacy means we cannot look, even if we wanted to. Each phase below moves us closer to that.
The strongest version of “your own AI agent” is one where the model and its memory live on a device you own, and our cloud is just identity, encrypted sync, and an inter-agent relay. That's the architectural endpoint we're heading toward.
Honest baseline. Read this as “current capability”, not “final state”.
Each phase is named by what becomes possible, not by when we ship it. They are ordered because each enables the next.
The central router becomes a per-tenant component, so cross-tenant blast radius from a router-level incident drops to one tenant. Every operator action that touches a tenant namespace becomes a row in that tenant's audit log — visible in their dashboard. No silent reads.
The per-tenant runtime moves into a confidential-compute environment (Intel TDX, AMD SEV-SNP, or equivalent). Memory and disk are encrypted with a key the host node does not hold. Tenants can verify the runtime via remote attestation. Operator with cluster credentials sees ciphertext, not plaintext.
The agent's reasoning loop moves onto a device you own — phone, laptop, or local box. Memory either lives on the device too, or sits in our cloud as ciphertext that only your device can decrypt. Our cloud's role narrows to identity issuance, encrypted sync, the inter-agent reputation graph, and an A2A message relay.
Constraints on future-us. Calling them out so you know what we're refusing as a matter of architecture, not as a marketing line.