Skip to content

research: LMEB-guided embedding model selection + domain fine-tuning for org memory #695

@Aureliolo

Description

@Aureliolo

Context

Two findings on embedding quality for agent memory:

  1. LMEB Benchmark -- 22 datasets, 193 tasks. MTEB performance does NOT generalize to memory retrieval (correlation ~-0.13). Episodic/dialogue/procedural taxonomy maps directly to SynthOrg's memory use cases.
  2. NVIDIA Domain-Specific Embedding Fine-Tune -- Automated pipeline (synthetic data gen, hard negative mining, contrastive fine-tuning). No manual annotation. Single GPU. +10-27% retrieval improvement.

Action Items

  • Evaluate current embedding model against LMEB leaderboard (not MTEB)
  • Select embedding model optimized for episodic + procedural memory retrieval patterns
  • Design optional embedding fine-tuning as OrgMemoryBackend initialization hook
  • Pipeline: synthetic data from org documents -> hard negative mining -> fine-tune -> deploy

References

Metadata

Metadata

Assignees

No one assigned

    Labels

    prio:highImportant, should be prioritizedscope:medium1-3 days of workspec:memoryDESIGN_SPEC Section 7 - Memory & Persistencetype:researchEvaluate options, make tech decisionsv0.6Minor version v0.6v0.6.0Patch release v0.6.0

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions