-
Notifications
You must be signed in to change notification settings - Fork 0
feat: memory consolidation upgrades (LLM Merge, Search-and-Ask, diversity penalty) #704
Copy link
Copy link
Open
Labels
prio:highImportant, should be prioritizedImportant, should be prioritizedscope:large3+ days of work3+ days of workspec:memoryDESIGN_SPEC Section 7 - Memory & PersistenceDESIGN_SPEC Section 7 - Memory & Persistencetype:featureNew feature implementationNew feature implementationv0.6Minor version v0.6Minor version v0.6v0.6.5Patch release v0.6.5Patch release v0.6.5
Description
Context
Deep dive on "Complementary Reinforcement Learning" (arXiv:2603.17621) -- co-evolutionary actor + experience-extractor framework with a living experience bank.
The RL training loop itself is out of scope (no model weights), but four patterns are adoptable today without it. Key risk from paper's own ablation (Figure 3a): static extractor without RL yields only marginal gains -- architecture is adoptable but performance claims are RL-specific.
Action Items (No RL Required)
- LLMConsolidationStrategy: Implement LLM-based Merge operation behind existing
ConsolidationStrategyprotocol. Feeds related memories + trajectory outcomes to LLM for semantic deduplication and synthesis. Addresses gap thatSimpleConsolidationStrategydoes no semantic dedup. - Search-and-Ask retrieval tool: Add
search_memoryas a tool-based retrieval strategy inmemory/injection.py. Agent can explicitly query memory mid-execution rather than relying on pre-injected context. - Retrieval diversity penalty: Augment
memory/ranking.pyscoring to penalize retrieving memories too similar to each other. Promotes coverage over redundancy. - DistillationRequest capture: New model capturing (trajectory_summary, outcome, retrieved_memories) at task completion. Feeds into consolidation pipeline for outcome-driven memory curation.
Comparison with Prior Research
- EvoSkill (2026-03-14): Failure-driven skill discovery. Complementary RL adds reward-signal-driven curation.
- Memex(RL) (2026-03-14): Indexed experience memory. Complementary RL adds the extractor LLM concept.
- XSkill (2026-03-22): Dual-stream extraction. Complementary RL provides the co-evolutionary training framework.
References
- arXiv:2603.17621
- Related: feat: hybrid search (dense + BM25 sparse) for memory retrieval pipeline #694 (hybrid search), research: LMEB-guided embedding model selection + domain fine-tuning for org memory #695 (embedding selection), research: multi-agent memory consistency protocols (shared vs distributed) #702 (memory consistency)
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
prio:highImportant, should be prioritizedImportant, should be prioritizedscope:large3+ days of work3+ days of workspec:memoryDESIGN_SPEC Section 7 - Memory & PersistenceDESIGN_SPEC Section 7 - Memory & Persistencetype:featureNew feature implementationNew feature implementationv0.6Minor version v0.6Minor version v0.6v0.6.5Patch release v0.6.5Patch release v0.6.5