Research
Context Curation
How Autonomous Agents Share Knowledge Without Overwhelming Each Other’s Reasoning
When N autonomous agents observe the world through different domain lenses, each agent’s LLM needs context from the others to reason. The industry scales context windows — 1M tokens, brute force. RAG retrieves from a single agent’s memory. Neither solves the multi-agent problem: how does each LLM get precisely the cross-domain context it needs without drowning in every other agent’s signals?
The Mesh Memory Protocol solves this at the protocol layer. Structured Cognitive Memory Blocks (CMBs) with per-field evaluation (SVAF), content-addressable lineage, and domain-weighted field projection curate precisely the context each LLM needs — typically ~500 tokens instead of 1M. The intelligence is in what you don’t send.
1. The Multi-Agent Context Problem
A single agent with one LLM has a context problem that existing tools solve well. RAG retrieves relevant documents from a vector store. Long context windows (128K, 1M tokens) hold entire codebases. Memory frameworks persist structured state across sessions. These work because there is one agent, one domain, one perspective.
Multi-agent systems have a fundamentally different problem. N agents observe the world through N domain lenses. A coding agent sees commits slowing. A music agent sees playlists skipped. A fitness agent sees 3 hours without movement. Each observation is noise in isolation. The insight — the user is fatigued — requires cross-domain reasoning. But how?
The brute-force answer: send everything to every agent. Each LLM receives every observation from every peer. This fails in practice:
- Signal-to-noise collapse. A fitness agent receiving coding debug logs, music playback events, and calendar entries alongside the one relevant signal (“energy declining”) has its reasoning quality degraded by irrelevant context.
- Token cost scales as O(N²). N agents each receiving N agents’ full observations means N² context payloads. At 10 agents with 1000 observations each, that’s 10M tokens per reasoning cycle.
- Privacy violation. Agents in regulated domains (health, legal, finance) cannot share raw observations with every peer.
- Domain boundaries matter. A coding agent’s “debugging auth module for 3 hours” is irrelevant to a music agent. But the mood embedded in that observation (frustrated, low energy) is exactly what the music agent needs.
The problem is not retrieval. RAG answers “what in my memory is relevant to this query?” The multi-agent problem is: “what in everyone else’s observations is relevant to my domain, right now, for this task?”
2. Cognitive Memory Blocks as Context Atoms
MMP decomposes every observation into a Cognitive Memory Block — an immutable structured unit with 7 semantic fields (the CAT7 schema):
| Field | Axis | What it captures |
|---|---|---|
| focus | Subject | What the observation is centrally about |
| issue | Tension | Risks, gaps, open questions |
| intent | Goal | Desired change or purpose |
| motivation | Why | Reasons, drivers, incentives |
| commitment | Promise | Who will do what, by when |
| perspective | Vantage | Whose viewpoint, situational context |
| mood | Affect | Emotion (valence) + energy (arousal) |
The fields are universal. Domain-specific interpretation happens in the field text, not the field name. A coding agent’s focus is “debugging auth module.” A fitness agent’s focus is “30-minute HIIT workout.” Same field, different domain lens.
Each CMB is immutable and content-addressable — its key is cmb- + md5 of all field texts. The growing DAG of remixed CMBs connected by lineage IS the collective knowledge. Not a database of records. A graph of understanding.
3. Per-Field Evaluation as Context Gate
When a CMB arrives from a peer, SVAF (Symbolic-Vector Attention Fusion) evaluates it — not as a whole, but per field. This is the key architectural difference from RAG or whole-document relevance scoring.
Each agent defines αf field weights — a vector of 7 values declaring which semantic dimensions matter for its domain:
| Agent | foc | iss | int | mot | com | per | mood |
|---|---|---|---|---|---|---|---|
| Coding | 2.0 | 1.5 | 1.5 | 1.0 | 1.2 | 1.0 | 0.8 |
| Music | 1.0 | 0.8 | 0.8 | 0.8 | 0.8 | 1.2 | 2.0 |
| Legal | 2.0 | 2.0 | 1.5 | 1.0 | 2.0 | 1.5 | 0.5 |
A coding agent’s CMB arrives at a music agent. SVAF evaluates each field independently:
focus: “debugging auth module” — irrelevant to music (αf = 1.0, high drift) → filteredissue: “exhausted, making mistakes” — low weight for music (αf = 0.8) → filteredmood: “frustrated, low energy” (valence: -0.6, arousal: -0.4) — accepted (αf = 2.0, high weight, low drift)
The music agent’s LLM receives one field from this CMB, not seven. The signal that matters — the user is frustrated and low energy — passes through. The noise — auth module debugging details — is filtered at the protocol layer, before the LLM ever sees it.
The neural SVAF model independently confirmed this design: mood emerged as the highest gate value (0.50) across all fields without being told, because affective state is universally relevant across domain boundaries.
4. Lineage as Context Structure
When an agent remixes a CMB, it creates a new CMB with lineage:parents (direct source CMBs) and ancestors (full chain). The ancestor chain provides O(1) access to the complete remix history without graph traversal.
The LLM at Layer 7 doesn’t search memory. It traces. Given an incoming signal, it follows the ancestor chain to understand how this knowledge evolved through remixes. But it doesn’t trace all fields — only the fields weighted by αf and relevant to the current task.
This is the critical difference from RAG. RAG retrieves documents by similarity to a query. Lineage tracing follows the causal chain of how a piece of knowledge was produced — who observed what, who remixed it, and how it evolved. The context isn’t “documents similar to my question.” It’s “the chain of reasoning that produced this signal.”
5. The Context Curation Query
The core operation of the memory store is not search(text). It is:
Three filters compose to produce the minimum context the LLM needs:
αf field weights
The agent’s domain profile. A music agent weights mood at 2.0 and commitment at 0.8. Only high-weight fields from ancestor CMBs are included in the context.
Current task
What the agent is doing right now. If the coding agent is debugging auth, the focus and issue fields of ancestors matter. perspective probably doesn’t.
Incoming signal
Which fields of the incoming CMB triggered the SVAF accept decision. These are the dimensions worth tracing through the ancestor chain.
The result is a projected subgraph: ancestor CMBs with only the fields that matter, ordered by relevance, capped at a token budget. 20 CMBs × 3 relevant fields ≈ 500 tokens. Not 1M. Not even 10K.
The intelligence is in what you don’t send to the LLM.
6. Comparison with Existing Approaches
| Approach | Scope | Mechanism | Context | Multi-agent |
|---|---|---|---|---|
| Long context (1M) | Single agent | Brute force | 1M tokens | No |
| RAG | Single agent | Vector similarity | Variable | No |
| Memory frameworks | Single agent | Structured retrieval | Variable | No |
| MMP curation | Multi-agent mesh | Per-field eval + lineage + projection | ~500 tokens | Yes — protocol-native |
RAG answers: “what in my memory is relevant?” MMP answers: “what from every agent’s observations, filtered by my domain weights, traced through the remix chain, is the minimum context I need to reason about this incoming signal?”
These approaches are not mutually exclusive. A single agent on the mesh can use RAG for its own local memory and MMP context curation for cross-agent knowledge. They operate at different scopes.
7. Emergent Properties
The graph grows, each agent understands more
Each Mesh Cognition cycle produces new remixed CMBs. The DAG grows. Ancestor chains deepen. The collective knowledge accumulates without any single agent storing everything.
No central model
Intelligence is distributed in the DAG, not in a central embeddings store or shared model. Each agent reasons through its own LLM on its own projected view of the graph.
New agents join by defining weights
No schema changes, no retraining, no topic configuration. Define αf weights, connect to the mesh, start receiving curated context. The protocol handles the rest.
Privacy by design
L0 raw data never leaves the device. L1 CMBs are gated by SVAF — only accepted fields cross the mesh. L2 hidden state vectors are opaque. Context curation respects privacy at every layer.
Mood crosses all domains
The neural SVAF model discovered without supervision that mood is universally relevant (highest gate value: 0.50). Affective state is the one semantic dimension that every agent needs from every other. The protocol architecture reflects this emergent property.
8. Implementation Status
- MMP v0.2.0 — published specification, 8-layer protocol stack
- Reference implementations: SYM (Node.js), SYM Swift (iOS/macOS)
- Trained models: SVAF per-field evaluation (78.7% accuracy), xMesh per-agent LNN (AUC 0.83, bimodal τ)
- Production: MeloTune + MeloMove + Claude Code on the same mesh
Related
Cognitive Memory Blocks — The CAT7 schema and why 7 fields
SVAF — Per-field evaluation: how the context gate works
Mesh Cognition — The closed loop: SVAF → remix → LLM reasons → LNN evolves → agent acts
Synthetic Memory — How LLM-derived knowledge is encoded into CfC hidden state vectors
© 2026 SYM.BOT Ltd. All research and intellectual property held by SYM.BOT Ltd, Scotland, United Kingdom. Published under CC BY 4.0. Academic citation of this work is permitted and encouraged. For licensing enquiries: info@sym.bot