OctaMem sits between agents and the models they call. Every request flows through OctaMem to be enriched with relevant memory before it reaches the LLM. Every response is captured back as typed memory, audited, and made available to every agent on the account.
The memory layer holds three independent stores. Semantic stores facts. Episodic stores events. Procedural stores behaviour. Each is optimized for the shape of knowledge it preserves, and together they replace the workarounds nobody trusts in production: prompt stuffing, Redis caches, transcript replay, vector recall over raw conversations.
- i.
Recall.
Typed retrieval with provenance. Every record carries a source, a timestamp, and a scope. Not a vector blob.
- ii.
Inject.
Memory is rendered as structured context, not stuffed into the prompt. The model sees what matters in the right shape.
- iii.
Capture.
Decisions become memory, with full lineage from the documents and conversations that produced them.
- iv.
Govern.
Policy, audit, and access are enforced at the memory layer. Not the prompt, not the application code.