Back to Arena
∞ Former
by Instituto de Telecomunicações / DeepMind / IST (Martins, Marinho, Martins)
System Card
OrganizationInstituto de Telecomunicações / DeepMind / IST (Martins, Marinho, Martins)
Released2021-09
Architectureexternal-memory-network / Continuous-space attention over long-term memory
DetailsContinuous-space long-term memory accessed via a continuous attention mechanism, so attention complexity is independent of context length. "Sticky memories" give more space to frequently accessed regions (biologically inspired).
Parameters—
Domainlong-context
Open SourceNo
PaperView Paper
acl-2022continuous-attentionsticky-memoriesunbounded
Capability Profile
Benchmark Scores
6 of 14 benchmarksMulti-Turn Recall0/2
LoCoMo
no dataMemoryBank
no dataCross-Session Memory0/1
LongMemEval
no dataMulti-Hop QA1/3
Agent Task Memory0/1
AgentBench-Mem
no dataPersonalization0/1
PerLTQA
no dataFactuality / Grounding0/1
RAGAS
no dataSources:∞ Former paper (arXiv:2109.00301); evaluated on BABILong: Testing the Limits of LLMs with Long-Context Reasoning-in-a-Haystack (AIRI, 2406)∞ Former paper (arXiv:2109.00301); evaluated on InfiniteBench: Extending Long Context Evaluation Beyond 100K Tokens (Tsinghua / OpenBMB, 2402)∞ Former paper (arXiv:2109.00301); evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)∞ Former paper (arXiv:2109.00301); evaluated on LooGLE: Can Long-Context Language Models Understand Long Contexts? (Peking University, 2311)∞ Former paper (arXiv:2109.00301); evaluated on Needle in a Haystack (Greg Kamradt, 2024)∞ Former paper (arXiv:2109.00301); evaluated on RULER: What's the Real Context Size of Your Long-Context Language Models (NVIDIA, 2404)