Back to Arena
Vespa AI
by Yahoo / Vespa.ai (independent OSS project)
System Card
OrganizationYahoo / Vespa.ai (independent OSS project)
Released2017-09
Architecturevector-rag / hybrid lexical-vector engine
DetailsVespa is a full-featured search and vector database engine supporting HNSW approximate nearest-neighbor search combined with BM25 lexical scoring in a single query. It runs as a stateful distributed service with multi-phase ranking, custom ranking expressions, and native ML inference. Designed for large-scale production workloads at Yahoo, it handles both batch updates and real-time writes.
Parameters—
Domainrag-retrieval
Open SourceYes
WebsiteVisit
hybrid-searchANNrankingreal-timedistributed
Capability Profile
Benchmark Scores
5 of 14 benchmarksMulti-Turn Recall0/2
LoCoMo
no dataMemoryBank
no dataCross-Session Memory0/1
LongMemEval
no dataMulti-Hop QA2/3
Agent Task Memory0/1
AgentBench-Mem
no dataPersonalization0/1
PerLTQA
no dataFactuality / Grounding1/1
Sources:Vespa AI vendor documentation; evaluated on HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering (Stanford / CMU, 1809)Vespa AI vendor documentation; evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)Vespa AI vendor documentation; evaluated on MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries (HKUST, 2401)Vespa AI vendor documentation; evaluated on RAGAS: Automated Evaluation of Retrieval-Augmented Generation (Exploding Gradients, 2309)Vespa AI vendor documentation; evaluated on RULER: What's the Real Context Size of Your Long-Context Language Models (NVIDIA, 2404)