eOS Continuum Welcome
An AI agent should work the way a colleague does: it remembers what it did yesterday, it reaches for tools it has used before, it picks up an interrupted task after a restart, it acts as one continuous identity over time. Today's platforms approximate that by stitching together a database for memory, a queue for durability, a vector store for recall, an identity layer for permissions, and application code that holds the layers in alignment. The approximation works until something fails between the layers.
The state collision
Agent platforms today are orchestration layers wrapped around stateless LLM calls. The LLM [[Substrate -- the runtime layer all application code and state run on|substrate]] is a function: input prompt, output tokens, no persistent state across calls. The agent's logical model is the opposite: I remember what I did, I have tools, I survive a restart, I am a continuous identity acting over time. [[The Agent's Logical Model|The agent's logical model]] and the LLM substrate's actual model are different categories of system, glued together by application code.
The mismatch is structural, not accidental. Most contemporary bugs in agent infrastructure are artifacts of the impedance: serialization boundaries between reasoning and stored memory, partial-failure ambiguity in multi-step plans, retry-versus-replay confusion when tool calls error mid-execution, vector-database-versus-ground-truth divergence, schema drift between what the LLM thinks it knows and what storage holds. These bugs do not get patched on the wrong substrate; they cease to be possible on a substrate that does not have them as failure modes.
Partial answers that do not converge
Each piece of state the agent needs has been engineered as a separate vendor: a vector database approximates memory, JSON schemas approximate tool definitions, queues approximate workflow durability, IAM middleware approximates capability bounds, bespoke retry logic approximates atomicity. Each vendor addresses one failure mode at one layer, then introduces a new failure mode at the boundary between itself and the next layer. The application code becomes the layer that holds the boundaries together -- and that layer is exactly where the bugs the previous section names live.
Two-plus years into the recent agent-platform build-out, the variety of approaches is widening rather than converging. The MemGPT lineage alone has produced A-MEM, Mem0, MemMachine, Memoria, and [[EverMemOS Self-Organizing Memory (2026)|EverMemOS]] in a span where each successor reports gains over the last. The successor pattern is itself the evidence: each new system inherits the previous one's substrate assumptions, names a new failure mode at the substrate boundary, and proposes a user-space mechanism to address it. The ceiling moves; the substrate does not.
The research is converging on a substrate-shaped need
The substrate-shaped requirement has now arrived from four independent research directions:
Prompt as environment. [[Recursive Language Models (Zhang et al., 2025)]] gives the LM a Python REPL in which the prompt is bound to a variable and the LM writes code that recursively calls itself over derived snippets. Intermediate state lives in named REPL variables; only constant-size metadata feeds back to the root LM. The architectural inversion is unmistakable -- the LM is a tool the program calls, not the program that calls the tools -- and the empirical result is concrete: inputs two orders of magnitude beyond model context windows at comparable cost. The paper names its own limitations in the same breath: sub-calls are blocking, recursion depth bounded, REPL state evaporates on process death. Its successor SRLM extends the scheme with uncertainty-aware trajectory selection and reports 22 percent accuracy improvement under identical wall-clock budgets. The direction is a microcosm of substrate-shaped agent reasoning at single-recursion scale -- and the limitations the paper itself names are an inventory of the substrate properties absent from Python.
Harness-managed virtual memory. ClawVM proposes a runtime layer inside the [[Harness -- the orchestration layer that drives an agent's LLM calls and tool use|harness]] that enforces atomicity (three-phase staged-update with scoped commit), persistence guarantees, and [[Capability -- a runtime-enforced authority naming what a piece of code may read, modify, or invoke|capability]] boundaries (pages carry scope metadata; writeback is checked against the active principal). The paper's load-bearing framing: "when a runtime manages a fast, scarce tier alongside slow, durable storage, the answer is virtual memory, not best-effort heuristics. Operating systems solved this decades ago." It eliminates all policy-controllable faults across budgets and reaches 100 percent task success against 76.7 percent for practitioner-configured baselines at tight budget. ClawVM is the closest external cousin to this graph's diagnosis, arrived at independently from the LLM-agents direction, and stops one architectural layer short of the substrate.
Agent memory as an OS concern. MemGPT framed agent memory as virtual-memory-style paging between context window (physical memory) and external storage (disk). The framing has run for over two years through the lineage named above; the gains compound, but the framing's natural conclusion -- if memory is an OS concern, atomicity is also an OS concern, and so are capability separation, multi-agent coherence, hot reload, sandboxed code load, asynchronous events, and state introspection -- sits in the literature without yet being claimed.
Compound AI systems. The [[Compound AI Systems (Zaharia et al., 2024)|Berkeley framing]] of orchestration over multiple LLMs and tools as a graph of operations is now surveyed in the optimization-survey literature, with runtime-shaped answers arriving at the substrate-vs-glue diagnostic from the orchestration direction. The Khattab axis -- DSPy as the user-space framework, DSPy Agent Skills as the practitioner extension, RLM as the inference pattern -- is one coherent program whose conclusions point at the same substrate. Prime Intellect frames RLM as "the paradigm of 2026" and explicitly targets long-horizon agent tasks spanning weeks to months, where orthogonal persistence stops being a stretch goal and becomes baseline.
Four directions, one diagnosis: agent infrastructure needs substrate properties the current substrate does not provide. The literature is naming the missing properties; the literature has not yet committed to providing them at the substrate layer.
The substrate diagnosis
This graph rests on the diagnostic the literature is converging on: [[Agent Runtimes Require Substrate Primitives, Not External Glue]]. The harness-layer and user-space answers do better than nothing -- ClawVM at the harness, MemGPT-and-successors in user-space, RLM at the recursion layer, Compound AI Systems at the orchestration layer all measurably move the ceiling. They share a structural property: each enforces substrate semantics on top of a substrate that does not natively carry them. The structural answer is to address the state problem at the layer where state lives -- in the runtime -- rather than to reconstruct it above.
That answer requires a runtime with one foundational property.
Orthogonal persistence
[[Orthogonal Persistence Is the Foundational Substrate Primitive|Orthogonal persistence]] is the property the diagnosis names. The application's entire state -- code, objects, in-flight operations, accumulated context -- lives in a single persistent image. State does not live in a database the runtime synchronizes with; the runtime is the state. Snapshot the runtime, restore it tomorrow, work continues from where it left off. There is no serialization boundary between reasoning and stored memory because there is no boundary; there is no synchronization protocol between application heap and external store because there is no external store.
The property is rare in contemporary infrastructure but not new. It has been the architectural floor of the [[The Capability-OS Tradition|capability-OS tradition]] (KeyKOS at Tymshare in the 1980s, EROS at Penn in the 1990s, CapROS, Coyotos), of Smalltalk and Lisp images, and of mainframe single-level-store designs. The category fell out of fashion when stateless services plus relational database became dominant in the 1990s. One member of the class -- descended from the same lineage, currently maintained by its original author -- has run production services for nearly three decades. What is new is shaping it for the contemporary agent-infrastructure audience.
The other seven runtime [[Primitive -- a property the runtime provides natively rather than as external infrastructure|primitives]] derive their value from orthogonal persistence. Atomic operations matter because they atomically affect persistent state. Capability bounds matter because the bounded namespace persists. Hot reload matters because the running state survives the reload. Sandboxed code load matters because the loaded code and any objects it creates are persistent artifacts, not ephemeral process state. Asynchronous events matter because they fire atomically with persistent state changes. Multi-agent coherence matters because shared state must survive across agents and restarts alike. State introspection matters because the queryable surface IS the persistent state graph. Without orthogonal persistence, the other seven are local-process concerns with no continuity across the agent's lifetime; with it, they are substrate-scope guarantees that survive restart indefinitely.
The eight primitives
-
[[Orthogonal Persistence Is the Foundational Substrate Primitive|Orthogonal persistence]] (foundational) -- every object, variable, and reference persists automatically across process restarts as a single durable image. The runtime owns the heap. Application code does not save or load.
-
[[Agent Operations Commit Wholly or Roll Back Wholly|Atomic operations]] -- every operation runs as a transaction over the in-memory state. If anything in the operation fails, the entire in-memory state reverts -- not just database rows, but every object that was created, mutated, or referenced during the failed attempt.
-
[[Capability Boundaries Are Runtime-Enforced, Not Policy-Checked|Capability separation]] -- each piece of code runs as a privileged identity with explicit capabilities defining what it can read, modify, or invoke. Out-of-capability operations do not return error codes; they fail at runtime and the operation's atomic boundary rolls back the partial work.
-
[[Multi-Agent Coherence Is Substrate-Serialized, Not Application-Coordinated|Coherent multi-agent semantics]] -- concurrent operations see the same state at the same logical time. The runtime serializes conflicting writes deterministically. There is no eventual consistency, no leader election, no distributed-lock dance.
-
[[Hot Reload Is a Runtime Operation, Not a Deployment Event|Hot logic update]] -- source code can be modified, recompiled, and merged into the live runtime while operations are in flight. In-progress operations finish with the old logic; subsequent operations use the new. No restart, no state loss, no warmup.
-
[[Code Load Compiles Into the Live Runtime, Bounded by Capability Tiers|Sandboxed code load]] -- new code can be authored at runtime, compiled by the runtime, and merged into the same persistent image as everything else. The capability layer enforces what newly-loaded code is allowed to do; the persistence layer ensures it stays loaded across restarts.
-
[[Event Notification Is Atomic With State Change, Not Polled or Queued|Asynchronous events]] -- when state changes that other code cares about, the runtime synchronously notifies registered listeners within the same atomic operation that caused the change. Subscribers do not poll, do not race, do not miss events because of delivery-versus-write ordering.
-
[[State Introspection Reads the Live State Graph, Not a Synthesized Mirror|State introspection]] -- the runtime's live state graph is directly queryable -- by application code, by humans, by other LLMs. The data model is the API.
The architectural inversion
Conventional pattern: an agent harness orchestrates LLM calls, treats LLMs as tools (alongside file read, web search, code execution), and rebuilds state, atomicity, identity, and event coordination from external infrastructure -- a database for state, a queue for atomicity, IAM middleware for identity, distributed locks for coordination, webhooks for notifications, deploy-time tooling for logic updates.
Inverted pattern: [[The Harness Drives the Runtime as a Tool, Alongside LLMs|the harness uses the runtime as a tool]], the same way it uses LLMs. The harness sends operations to the runtime; the runtime returns state and events. State, mutation safety, multi-agent coordination, identity boundaries, and event notification all arrive as the runtime's natural output.
| Property | LLM-as-tool | Runtime-as-tool |
|---|---|---|
| State across calls | None | Persistent shared graph |
| Mutation safety | None | Atomic call-tree rollback |
| Multi-agent coordination | External | Coherent runtime enforces |
| Identity boundaries | None | Capability-separated |
| Event notifications | None | Native asynchronous |
| Context window | Fixed tokens | Arbitrary state, queryable as needed |
| Observability | Opaque | Object introspection, event log |
LLMs and the runtime sit at the same architectural level in the harness's tool surface, distinguished by cadence and primitive set rather than by rank. The harness's authority over orchestration is preserved; what the runtime adds is the substrate-primitive set LLMs structurally cannot provide. RLM proves the inversion at single-recursion scale on Python; the substrate proves it at runtime scale on a substrate authored for the role.
Five-axis containment for LLM-authored code
LLMs make mistakes. Code an agent (or an LLM) loads into the runtime sits inside [[Agent Code Containment Stacks Five Axes, Not One|five independent containment mechanisms]], each bounding a different attack surface:
-
Language constraint. The DSL agents write in cannot express dangerous operations -- no raw memory access, no escape to the host language, no direct system-call surface.
-
Location constraint. Agent code lives only in agent-owned objects, never in privileged runtime layers. The kernel and system tiers are inaccessible regardless of what the language permits.
-
Invocation constraint. Agent code is only entered through the event system, not by direct calls from arbitrary code. An attacker who has not subscribed to events cannot reach the code.
-
Capability separation. What agent code can reach when it does run is mediated by the privilege layer.
-
Atomic rollback. Any error in agent code reverts the entire operation. No partially-applied state, no manual cleanup, no half-corrupted state for the next agent to reconcile.
To break out, agent code would have to defeat all five mechanisms simultaneously. Stacked containment changes the cost calculus: defeating language puts the attacker in a location they cannot run from; defeating location puts them in front of an invocation constraint they cannot enter through; defeating invocation puts them inside a capability layer that bounds reach; defeating capability still leaves them inside an atomic envelope that rolls back partial state.
| Platform | Axes covered |
|---|---|
| WebAssembly + WASI | 2 (sandbox + capability) |
| Lua embedded in game engines | 1 (language) |
| Erlang processes | 2 (location + supervision-restart) |
| This substrate | 5 |
The 5/5 claim is grounded in [[Meriadoc (ChatTheatre, 2026)]], the reference implementation's sandboxed code-load DSL -- a substrate-layer mechanism in continuous production for over twenty-five years, with each axis enforced by a specific source-code mechanism a reviewer can verify directly.
For LLM-generated code specifically, this is the structural safety the workload requires. The runtime contains the mistakes -- not the developer, not the prompt-injection detector, not the human-in-the-loop reviewer.
The lineage that quietly carried the answer
The architecture that carries all eight primitives together has run production services for nearly three decades. It powered iChat -- the technology behind the [[iChat Production Pedigree|first Yahoo! Chatrooms]] in the late 1990s. Production deployments in the same lineage have run continuously for over twenty years. The codebase is open-source, actively maintained by its original author, and current as of 2026.
[[The Capability-OS Tradition]] -- KeyKOS in the 1980s at Tymshare, EROS in the 1990s at Penn, CapROS, Coyotos -- plus Smalltalk and Lisp images, plus mainframe single-level-store designs -- are the architectural ancestors. These systems explicitly used [[Orthogonal Persistence as Foundational Primitive|orthogonal persistence]] as the architectural property. The whole category fell out of fashion when stateless services + relational database became dominant in the 1990s; the recent agent-infrastructure literature is the field independently re-deriving the requirements that fall-out left abandoned.
The active member of the class is [[DGD as the Living Member of the Lineage|Felix Croes's DGD]] (the Dworkin Game Driver), the runtime behind iChat, Castle Marrach, and other long-running narrative-MUD productions. Twenty-five-plus years of operational soak on a [[Adopt Single-Coherence-Domain Architecture|single coherence domain]], with the same architectural primitives modern agent infrastructure is being asked to provide. eOS Continuum's approach is to take the lineage's substrate and shape it for the contemporary agent-infrastructure audience, rather than reinventing the substrate from scratch on a stack that does not carry orthogonal persistence as a property. The first iteration is being built in [[eos-harness Minimum Viable Architecture|eos-harness]] (a sibling repository), which validates the substrate's behavior end-to-end before the broader surface is built.
What becomes possible
Three shapes of customer problem the substrate's primitives change the engineering math for. None is hypothetical; the substrate has run shapes structurally similar to these for over two decades.
[[Customer-Authored Automation in a SaaS Product]]. Customers want to extend a SaaS product with real logic -- not just webhooks, not just expression-language formulas, but their own rules, calculations, integrations, conditional workflows. Today: ship a limited expression language (Notion formulas), let them write per-tenant cloud functions (operational complexity), or accept that extensibility is the customer-driving feature you keep deferring. With this substrate: customer-authored code compiles into a capability-bounded namespace inside the runtime; the namespace persists across restarts; isolation is a primitive, not a Docker tax. Salesforce Apex without the Salesforce, Notion formulas with teeth.
[[Long-Running Stateful Workflows]]. Order fulfillment, loan applications, hiring pipelines, multi-step customer onboarding -- processes that run for days or weeks across many possible paths. Today: pick Temporal, AWS Step Functions, or roll a workflow engine on database + queue + retries; reconcile inconsistent half-completed processes when something fails mid-stream. With this substrate: the workflow IS the persistent runtime state. Each step runs atomically; failure rolls back, no half-completed processes leaving inconsistent records. Hot reload fixes bugs in workflows that are mid-flight without losing the customer's accumulated progress.
[[AI-Authored Tools and Durable Agent Memory]]. Customers want AI to do real work on their data -- generate reports, refactor code, run analyses, drive multi-step actions across their systems. The agent has to write and execute code as part of its reasoning, and it has to remember what it learned. Today: limited tool calls, slow Docker-isolated sandboxes that warm up per session, weak isolation if you go local, agent state that resets every conversation. With this substrate: agent-authored code compiles into a capability-bounded namespace at near-native speed, persists for the agent to reuse across sessions, and cannot escape its boundary. The customer experience changes from "the AI ran for thirty seconds in a sandboxed timeout" to "the AI built me a custom tool that's still here next week."
Where this fits in the agent ecosystem
This is a substrate, not an agent framework. It [[The Substrate Composes Additively With the Agent Ecosystem|composes with]] the tools the audience already uses:
- LangChain / LlamaIndex / DSPy / similar reasoning frameworks -- the substrate is the runtime your agents run on top of. These frameworks remain useful for prompt-shape and reasoning-pattern abstraction.
- MCP / A2A / agent-to-agent protocols -- the substrate's HTTP and event primitives implement these protocols natively. The runtime can be both an MCP server (exposing tools to external agents) and an MCP client (calling external tools from inside).
- Vector databases (Pinecone, Weaviate, Chroma, pgvector) -- still useful for fuzzy retrieval over large unstructured corpora. The substrate handles exact memory of structured agent state. Both can coexist.
- Workflow orchestrators (Temporal, AWS Step Functions, Inngest) -- the substrate handles small-to-medium-scale durable workflows natively. Temporal remains the right tool for cross-machine workflow durability at large scale.
- Sandboxed code execution (E2B, Modal, Replit Agents, Daytona) -- the substrate makes sandboxed code-execution a runtime primitive rather than a per-call infrastructure choice. For one-off cloud-scale code execution, those tools remain appropriate.
The composition story is deliberately additive: the runtime fills the substrate gap that current platforms paper over, without asking adopters to rewrite the parts that already work. Migration is incremental -- start with one workflow, validate the substrate's behavior, add more.
Where this is wrong
Use a different substrate if any of the following apply:
- [[Stateless Workloads Do Not Need This Substrate|Stateless workloads]]. Pure RAG queries -- input prompt, retrieve, answer, forget -- get no benefit from persistence overhead. A stateless function-calling pattern on top of an LLM provider is the right choice.
- [[Embarrassingly Parallel Inference]]. Cloud-scale parallel LLM calls (batch inference over millions of documents) want a different shape than a single coherence domain offers.
- [[Five-Nines Availability Workloads]]. Single-machine architecture inherits single-machine failure modes. Multi-region active-active distributed systems are the right substrate for that requirement; they trade the atomicity guarantees this runtime provides for the availability they need.
- [[Workloads That Need Distributed Coherence]]. Global ledgers, planet-scale chat, content delivery -- the substrate's [[Adopt Single-Coherence-Domain Architecture|single-coherence-domain]] commitment is the wrong fit at the architecture level.
- [[Existing-Stack Inertia]]. Teams deep in existing orchestration code whose current platform's pain is not severe -- wait until the substrate pain is felt before paying migration cost.
- [[Managed-Cloud-Only Operational Models]]. Self-hosting is first-class here; managed-service offerings are downstream.
The structural rule: [[The Substrate Is the Wrong Choice Without Substrate-Pain|the substrate is for workloads where the primitives reduce real complexity in your application code]]. If the application does not fight the substrate problems the runtime solves, the substrate adds overhead without benefit.
The demonstration
The eOS Continuum project's demonstration target is [[Passkey-Protected Sandboxed Web REPL|a passkey-protected sandboxed Web REPL]] -- five minutes in a browser tab to see what the substrate carries.
An authenticated user (or agent, or human acting on behalf of an agent) authenticates via Passkey -- Touch ID, Face ID, or hardware key. The browser drops them into a code editor running against a bounded namespace inside the runtime: their workspace, where they build tools, materialize objects, accumulate state. The capability layer prevents the workspace's code from touching anyone else's namespace, the runtime kernel, or the host operating system.
Kill the runtime process. Restart it. Log back in via Passkey. The workspace is exactly as it was left. Every tool built, every object materialized, every accumulated piece of context, intact -- not because anything was serialized to a database between sessions, but because the runtime is the state and orthogonal persistence keeps the entire image durable across restarts.
Seven of the eight primitives are exercised by this single artifact: orthogonal persistence (kill/restart preserves state), atomic operations (each REPL evaluation as an atomic envelope), capability separation (bounded namespace), coherent multi-agent semantics (multi-tenant Web REPL), hot reload (REPL recompiles without restart), sandboxed code load (the REPL's own evaluator), state introspection (queryable namespace). Only asynchronous events is unrepresented, and that primitive lights up as soon as multiple workspaces interact.
License, sovereignty, self-hosting
The license posture is part of the position, not an afterthought:
- [[Adopt BSD-2-Clause-Patent License|Permissively licensed]] where the team has freedom to choose. Code the team writes from scratch is suitable for downstream use, including in commercial proprietary products.
- Self-hostable as a primary deployment model. Run the runtime on your own hardware, in your own cloud account, behind your own network. No mandatory cloud-side dependency, no telemetry-by-default, no per-call usage fees.
- Auditable. The substrate is small enough that one careful engineer can read and understand it. The kernel layer that enforces capability boundaries is the security-critical surface; it is intentionally minimal.
- Forkable. If the team's direction diverges from yours, fork. If the project goes dormant, fork. The license and architecture together mean your agent infrastructure is not dependent on the team's continued existence.
- [[Data Exit Is a First-Class Feature|Data exit is a first-class feature]]. Snapshots are documented, versioned, and self-describing. Tools to export the state graph in a portable format ship with the runtime.
These commitments reflect a posture: the runtime is infrastructure for builders, and builders need [[Builders Own the Substrate They Depend On|sovereignty]] over their infrastructure to commit to it for serious work.
What this site is for
A common reference for the small team and early collaborators. A starting point for the documentation introduction, the technical posts, the conversations with future contributors. The argument is intentionally framed for the small-team / community / audience-of-builders model.
What this site is, affirmatively:
- A reference for the small team.
- A starting point for collaboration with the contemporary agent-infrastructure audience.
- A test of the architectural claim. If a sophisticated builder reads this and the value proposition lands, the architecture is right. If they shrug, either the audience targeting is wrong or the substrate is genuinely not different enough -- and that is information worth getting early.
Entry Points
To browse every node in the graph at once, see the full node directory — every node grouped by form type and listed alphabetically, with grafted nodes marked so the donor's source is one click away.
Where to go from here, depending on what brought you:
- Curious about the diagnostic. The "state collision," "partial answers that do not converge," and "research is converging" sections above are the diagnosis in compressed form. [[Agent Runtimes Require Substrate Primitives, Not External Glue]] when authored will carry the load-bearing argument as a Conviction node; until then, the agentic-runtime concept brief in [[eos-harness Minimum Viable Architecture|eos-harness]] carries the longer-form version.
- Curious about the eight primitives. Each will be authored as a Conviction node under
nodes/Convictions/. They are listed above with ghost links; the per-primitive Conviction texts already live in [[eos-harness Minimum Viable Architecture|eos-harness]]'s graph and will migrate or scion here as the work proceeds. - Evaluating against your workload. The "Where this is wrong" section is the fast filter -- if your workload matches one of the six cases there, save yourself the further reading. Otherwise the [[Builders Own the Substrate They Depend On|sovereignty]] section names what commitment looks like.
- Skeptical about competing with matched-pair vendors. The empirical case for agent harness fit as a competitive moat for incumbents is real, and eOS Continuum does not natively have it. The project's load-bearing answer is [[Substrate Compounding Overcomes Agent Harness Fit]] -- on long-horizon stateful workloads, the substrate's compounding outweighs the matched pair's per-turn quality. The Conviction's drift-recognition section names the failure modes for this bet.
- Considering contributing. Read
AGENTS.mdfor the agent-orientation framing and graph conventions, thennodes/Contracts/for the structural rules each node form satisfies, thennodes/Skills/for the agent-invocable workflows. The active implementation work happens in [[eos-harness Minimum Viable Architecture|eos-harness]]; this graph is the architectural argument that work tests.
The graph is pre-content as of this writing. Most of the wikilinks above are ghost links to forthcoming nodes -- each represents a claim the project is committed to articulating in its own typed-node form. The pace of authoring is set by the work in [[eos-harness Minimum Viable Architecture|eos-harness]] and by the conversations with early collaborators that the small-team-of-builders model depends on.
Out of Scope
What this Touch Point's lens deliberately does not address:
- Marketing. No product naming sales pitch. A separate document would be required for any commercial framing.
- Fundraising. A separate document would be required for an investor audience; the argument here is intentionally framed for builders, not capital allocators.
- Technical specification. Primitives are described at the level of "what they do and why they matter," not "what the API looks like." Structural rules live in the Form Contracts under nodes/Contracts/.
- Product requirements. The work being done is architectural-depth-over-feature-breadth; product-requirements documents belong to a downstream phase.
- Competitive analysis. No specific platforms are named as adversaries. The argument is structural; any platform built on the wrong substrate has the same weaknesses, regardless of branding.
Relations
-
conforms_to::[[Touch Point Form Contract]]
- This Touch Point declares compliance with the Touch Point Form Contract — the form for nodes whose role is to frame a reader's lens onto a region of the graph.
-
frames_lens_on::[[Agent Runtimes Require Substrate Primitives, Not External Glue]]
- The load-bearing Conviction this graph anchors. Currently a ghost link; the Conviction will be authored as the substrate argument matures from the eos-harness Minimum Viable Architecture work into a typed Conviction node here.
-
frames_lens_on::[[eOS Continuum]]
- The Practice Domain Gloss this Touch Point orients first-time readers toward. The Gloss is currently a ghost link awaiting authoring; its role is to name the community of practice this graph speaks within.
-
composes_with::[[DeepContext.com Graph (Allen, 2026)]]
- The donor Reference proxying the meta-layer source. The Form Contract this Touch Point conforms to was grafted from there; concept Glosses, Predicates, and generic Skills referenced in this graph resolve there via external wikilinks. This Touch Point's content is eOS Continuum's own — locally authored, not grafted.