Research

Context Curation

How Autonomous Agents Share Knowledge Without Overwhelming Each Other’s Reasoning

Hongwei Xu·SYM.BOT Ltd·28 March 2026

When N autonomous agents observe the world through different domain lenses, each agent’s LLM needs context from the others to reason. The industry scales context windows — 1M tokens, brute force. RAG retrieves from a single agent’s memory. Neither solves the multi-agent problem: how does each LLM get precisely the cross-domain context it needs without drowning in every other agent’s signals?

The Mesh Memory Protocol solves this at the protocol layer. Structured Cognitive Memory Blocks (CMBs) with per-field evaluation (SVAF), content-addressable lineage, and domain-weighted field projection curate precisely the context each LLM needs — typically ~500 tokens instead of 1M. The intelligence is in what you don’t send.

1. The Multi-Agent Context Problem

A single agent with one LLM has a context problem that existing tools solve well. RAG retrieves relevant documents from a vector store. Long context windows (128K, 1M tokens) hold entire codebases. Memory frameworks persist structured state across sessions. These work because there is one agent, one domain, one perspective.

Multi-agent systems have a fundamentally different problem. N agents observe the world through N domain lenses. A coding agent sees commits slowing. A music agent sees playlists skipped. A fitness agent sees 3 hours without movement. Each observation is noise in isolation. The insight — the user is fatigued — requires cross-domain reasoning. But how?

The brute-force answer: send everything to every agent. Each LLM receives every observation from every peer. This fails in practice:

  • Signal-to-noise collapse. A fitness agent receiving coding debug logs, music playback events, and calendar entries alongside the one relevant signal (“energy declining”) has its reasoning quality degraded by irrelevant context.
  • Token cost scales as O(N²). N agents each receiving N agents’ full observations means N² context payloads. At 10 agents with 1000 observations each, that’s 10M tokens per reasoning cycle.
  • Privacy violation. Agents in regulated domains (health, legal, finance) cannot share raw observations with every peer.
  • Domain boundaries matter. A coding agent’s “debugging auth module for 3 hours” is irrelevant to a music agent. But the mood embedded in that observation (frustrated, low energy) is exactly what the music agent needs.

The problem is not retrieval. RAG answers “what in my memory is relevant to this query?” The multi-agent problem is: “what in everyone else’s observations is relevant to my domain, right now, for this task?”

2. Cognitive Memory Blocks as Context Atoms

MMP decomposes every observation into a Cognitive Memory Block — an immutable structured unit with 7 semantic fields (the CAT7 schema):

FieldAxisWhat it captures
focusSubjectWhat the observation is centrally about
issueTensionRisks, gaps, open questions
intentGoalDesired change or purpose
motivationWhyReasons, drivers, incentives
commitmentPromiseWho will do what, by when
perspectiveVantageWhose viewpoint, situational context
moodAffectEmotion (valence) + energy (arousal)

The fields are universal. Domain-specific interpretation happens in the field text, not the field name. A coding agent’s focus is “debugging auth module.” A fitness agent’s focus is “30-minute HIIT workout.” Same field, different domain lens.

Each CMB is immutable and content-addressable — its key is cmb- + md5 of all field texts. The growing DAG of remixed CMBs connected by lineage IS the collective knowledge. Not a database of records. A graph of understanding.

3. Per-Field Evaluation as Context Gate

When a CMB arrives from a peer, SVAF (Symbolic-Vector Attention Fusion) evaluates it — not as a whole, but per field. This is the key architectural difference from RAG or whole-document relevance scoring.

Each agent defines αf field weights — a vector of 7 values declaring which semantic dimensions matter for its domain:

Agentfocissintmotcompermood
Coding2.01.51.51.01.21.00.8
Music1.00.80.80.80.81.22.0
Legal2.02.01.51.02.01.50.5

A coding agent’s CMB arrives at a music agent. SVAF evaluates each field independently:

  • focus: “debugging auth module” — irrelevant to music (αf = 1.0, high drift) → filtered
  • issue: “exhausted, making mistakes” — low weight for music (αf = 0.8) → filtered
  • mood: “frustrated, low energy” (valence: -0.6, arousal: -0.4) — acceptedf = 2.0, high weight, low drift)

The music agent’s LLM receives one field from this CMB, not seven. The signal that matters — the user is frustrated and low energy — passes through. The noise — auth module debugging details — is filtered at the protocol layer, before the LLM ever sees it.

The neural SVAF model independently confirmed this design: mood emerged as the highest gate value (0.50) across all fields without being told, because affective state is universally relevant across domain boundaries.

4. Lineage as Context Structure

When an agent remixes a CMB, it creates a new CMB with lineage:parents (direct source CMBs) and ancestors (full chain). The ancestor chain provides O(1) access to the complete remix history without graph traversal.

The LLM at Layer 7 doesn’t search memory. It traces. Given an incoming signal, it follows the ancestor chain to understand how this knowledge evolved through remixes. But it doesn’t trace all fields — only the fields weighted by αf and relevant to the current task.

This is the critical difference from RAG. RAG retrieves documents by similarity to a query. Lineage tracing follows the causal chain of how a piece of knowledge was produced — who observed what, who remixed it, and how it evolved. The context isn’t “documents similar to my question.” It’s “the chain of reasoning that produced this signal.”

5. The Context Curation Query

The core operation of the memory store is not search(text). It is:

curate(incomingCMB, agentWeights, currentTask) → contextForLLM

Three filters compose to produce the minimum context the LLM needs:

αf field weights

The agent’s domain profile. A music agent weights mood at 2.0 and commitment at 0.8. Only high-weight fields from ancestor CMBs are included in the context.

Current task

What the agent is doing right now. If the coding agent is debugging auth, the focus and issue fields of ancestors matter. perspective probably doesn’t.

Incoming signal

Which fields of the incoming CMB triggered the SVAF accept decision. These are the dimensions worth tracing through the ancestor chain.

The result is a projected subgraph: ancestor CMBs with only the fields that matter, ordered by relevance, capped at a token budget. 20 CMBs × 3 relevant fields ≈ 500 tokens. Not 1M. Not even 10K.

The intelligence is in what you don’t send to the LLM.

6. Comparison with Existing Approaches

ApproachScopeMechanismContextMulti-agent
Long context (1M)Single agentBrute force1M tokensNo
RAGSingle agentVector similarityVariableNo
Memory frameworksSingle agentStructured retrievalVariableNo
MMP curationMulti-agent meshPer-field eval + lineage + projection~500 tokensYes — protocol-native

RAG answers: “what in my memory is relevant?” MMP answers: “what from every agent’s observations, filtered by my domain weights, traced through the remix chain, is the minimum context I need to reason about this incoming signal?”

These approaches are not mutually exclusive. A single agent on the mesh can use RAG for its own local memory and MMP context curation for cross-agent knowledge. They operate at different scopes.

7. Emergent Properties

The graph grows, each agent understands more

Each Mesh Cognition cycle produces new remixed CMBs. The DAG grows. Ancestor chains deepen. The collective knowledge accumulates without any single agent storing everything.

No central model

Intelligence is distributed in the DAG, not in a central embeddings store or shared model. Each agent reasons through its own LLM on its own projected view of the graph.

New agents join by defining weights

No schema changes, no retraining, no topic configuration. Define αf weights, connect to the mesh, start receiving curated context. The protocol handles the rest.

Privacy by design

L0 raw data never leaves the device. L1 CMBs are gated by SVAF — only accepted fields cross the mesh. L2 hidden state vectors are opaque. Context curation respects privacy at every layer.

Mood crosses all domains

The neural SVAF model discovered without supervision that mood is universally relevant (highest gate value: 0.50). Affective state is the one semantic dimension that every agent needs from every other. The protocol architecture reflects this emergent property.

8. Implementation Status

  • MMP v0.2.0 — published specification, 8-layer protocol stack
  • Reference implementations: SYM (Node.js), SYM Swift (iOS/macOS)
  • Trained models: SVAF per-field evaluation (78.7% accuracy), xMesh per-agent LNN (AUC 0.83, bimodal τ)
  • Production: MeloTune + MeloMove + Claude Code on the same mesh

Related

Cognitive Memory Blocks — The CAT7 schema and why 7 fields

SVAF — Per-field evaluation: how the context gate works

Mesh Cognition — The closed loop: SVAF → remix → LLM reasons → LNN evolves → agent acts

Synthetic Memory — How LLM-derived knowledge is encoded into CfC hidden state vectors

© 2026 SYM.BOT Ltd. All research and intellectual property held by SYM.BOT Ltd, Scotland, United Kingdom. Published under CC BY 4.0. Academic citation of this work is permitted and encouraged. For licensing enquiries: info@sym.bot