Back to Arena
Recurrent Memory Transformer
by MIPT / DeepPavlov (Bulatov, Kuratov, Burtsev)
System Card
OrganizationMIPT / DeepPavlov (Bulatov, Kuratov, Burtsev)
Released2022-07
Architectureexternal-memory-network / Memory tokens passed between segments recurrently
DetailsAdds special memory tokens to each segment that pass information recurrently across segments of a long sequence, with no architectural changes beyond token-level memory slots.
Parameters—
Domainlong-context
Open SourceYes
PaperView Paper
CodeRepository
neurips-2022aaai-2024recurrentmemory-tokens1m-tokens
Capability Profile
Benchmark Scores
6 of 14 benchmarksMulti-Turn Recall0/2
LoCoMo
no dataMemoryBank
no dataCross-Session Memory0/1
LongMemEval
no dataMulti-Hop QA1/3
Agent Task Memory0/1
AgentBench-Mem
no dataPersonalization0/1
PerLTQA
no dataFactuality / Grounding0/1
RAGAS
no dataSources:Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on BABILong: Testing the Limits of LLMs with Long-Context Reasoning-in-a-Haystack (AIRI, 2406)Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on InfiniteBench: Extending Long Context Evaluation Beyond 100K Tokens (Tsinghua / OpenBMB, 2402)Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on LooGLE: Can Long-Context Language Models Understand Long Contexts? (Peking University, 2311)Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on Needle in a Haystack (Greg Kamradt, 2024)Recurrent Memory Transformer paper (arXiv:2207.06881); evaluated on RULER: What's the Real Context Size of Your Long-Context Language Models (NVIDIA, 2404)