Skip to content

research(memory): MemCoT test-time memory scaling — multi-view long-context reasoning with episodic + semantic short-term memory #3564

@bug-ops

Description

@bug-ops

Description

arXiv:2604.08216 (April 9, 2026) — MemCoT: Test-Time Scaling through Memory-Driven Chain-of-Thought.

MemCoT is a plug-and-play, training-free framework that addresses LLM hallucination and catastrophic forgetting during causal reasoning over massive, fragmented long contexts. Existing memory mechanisms treat retrieval as a static, single-step passive matching process, leading to semantic dilution and contextual fragmentation.

Core Components

Multi-view Long-Term Memory Perception Module:

  • Zoom-In: evidence localization — identifies where relevant evidence resides in the long context
  • Zoom-Out: contextual expansion — reconstructs surrounding causal structure necessary for reasoning

Task-Conditioned Dual Short-Term Memory:

  • Semantic state memory: tracks conceptual state across reasoning steps
  • Episodic trajectory memory: tracks the reasoning path itself (what was inferred, in what order)

Results

On the LoCoMo benchmark (long-context conversation memory), MemCoT achieves SOTA:

  • GPT-4o-mini with MemCoT: F₁ = 58.03 (vs baseline ~30)
  • Qwen2.5-14B with MemCoT: F₁ = 57.06

Compact models outperform much larger baselines on long-context memory tasks.

Relevance to Zeph

Zeph's memory system (MAGMA graph + SYNAPSE spreading activation + scene consolidation) is strong at structural/associative retrieval but doesn't address the reasoning dimension over long contexts. MemCoT's dual short-term memory maps naturally to Zeph's context builder:

  • Semantic state memory ↔ extend TurnContext with a mutable "semantic state" accumulator updated after each LLM turn
  • Episodic trajectory memory ↔ the reasoning-bank in ReasoningMemory already stores trajectories; MemCoT's Zoom-In/Out could feed the retrieval step
  • Multi-view LTM perceptionMemoryOrchestrator::recall could be extended with a zoom-in pass (localize) before the zoom-out pass (expand context)

This is training-free — it operates entirely at inference time through prompt engineering + retrieval structuring.

Proposed Design Direction

  • Add a SemanticStateAccumulator component to TurnContext that summarizes the "conceptual state" after each turn
  • Extend ReasoningMemory retrieval with Zoom-In (targeted evidence localization) before the existing spreading activation
  • Expose a memory.memcot.enabled config flag; default off; benchmark improvement on multi-turn factual tasks

References

Metadata

Metadata

Assignees

Labels

P3Research — medium-high complexityllmzeph-llm crate (Ollama, Claude)memoryzeph-memory crate (SQLite)researchResearch-driven improvement

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions