Back to Arena
Mnemosyne
by Johns Hopkins / independent (2025)
System Card
OrganizationJohns Hopkins / independent (2025)
Released2025-10
Architecturegraph-rag / Graph memory with human-inspired decay, refresh, and core-summary
DetailsGraph-structured LTM with modular substance/redundancy filters, commit and pruning, probabilistic recall using temporal decay and refresh. Adds a concentrated "core summary" derived from a fixed-length memory-graph subset.
Parameters—
Domainpersonalizationlifelong-learningepisodic-session
Open SourceNo
PaperView Paper
edgedecay-refreshgraphunsupervised
Capability Profile
Benchmark Scores
6 of 14 benchmarksLong-Context Retrieval0/5
RULER
no dataNIAH
no dataLooGLE
no dataLongBench
no data∞Bench
no dataMulti-Turn Recall2/2
Cross-Session Memory1/1
Multi-Hop QA1/3
Agent Task Memory1/1
Personalization1/1
Factuality / Grounding0/1
RAGAS
no dataSources:arXiv:2510.08601 Table 2 — Same paper as mnemosyne — paper is titled 'for Edge-Based LLMs'; edge is the main design targetMnemosyne paper (arXiv:2510.08601); evaluated on LongMemEval: Benchmarking Chat Assistants on Long-Term Interactive Memory (Salesforce AI Research, 2410)Mnemosyne paper (arXiv:2510.08601); evaluated on MemoryBank: Enhancing LLMs with Long-Term Memory (Sun Yat-sen University, 2305)Mnemosyne paper (arXiv:2510.08601); evaluated on PerLTQA: A Personal Long-Term Memory Question Answering Dataset (PolyU, 2402)Mnemosyne paper (arXiv:2510.08601); evaluated on AgentBench Memory Track (Tsinghua KEG, 2308)Mnemosyne paper (arXiv:2510.08601); evaluated on BABILong: Testing the Limits of LLMs with Long-Context Reasoning-in-a-Haystack (AIRI, 2406)