Back to Arena
REALM
by Google Research (Guu et al.)
System Card
OrganizationGoogle Research (Guu et al.)
Released2020-02
Architecturevector-rag / Latent retriever pretrained with MLM backprop
DetailsAugments LM pretraining with a latent knowledge retriever that attends over millions of documents. Retriever is trained unsupervised by masked-LM loss backpropagating through the retrieval step.
Parameters—
Domainrag-retrievalknowledge-graph
Open SourceYes
PaperView Paper
CodeRepository
icml-2020pretraininglatent-retrieverwikipedia
Capability Profile
Benchmark Scores
6 of 14 benchmarksMulti-Turn Recall0/2
LoCoMo
no dataMemoryBank
no dataCross-Session Memory1/1
Multi-Hop QA2/3
Agent Task Memory0/1
AgentBench-Mem
no dataPersonalization0/1
PerLTQA
no dataFactuality / Grounding1/1
Sources:REALM paper (arXiv:2002.08909); evaluated on HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering (Stanford / CMU, 1809)REALM paper (arXiv:2002.08909); evaluated on MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries (HKUST, 2401)REALM paper (arXiv:2002.08909); evaluated on RAGAS: Automated Evaluation of Retrieval-Augmented Generation (Exploding Gradients, 2309)REALM paper (arXiv:2002.08909); evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)REALM paper (arXiv:2002.08909); evaluated on LongMemEval: Benchmarking Chat Assistants on Long-Term Interactive Memory (Salesforce AI Research, 2410)REALM paper (arXiv:2002.08909); evaluated on RULER: What's the Real Context Size of Your Long-Context Language Models (NVIDIA, 2404)