Back to Arena
Galileo AI
by Galileo Technologies Inc.
System Card
OrganizationGalileo Technologies Inc.
Released2021-06
Architectureagentic-workflow / LLM evaluation with Luna foundation model
DetailsGalileo's core differentiator is Luna, a family of compact Evaluation Foundation Models fine-tuned specifically for hallucination detection, toxicity, prompt security, and data leak detection. ChainPoll methodology achieves 85% correlation with human feedback. Luna replaces expensive LLM-as-judge calls with low-latency, low-cost specialized models. Also offers guardrails, data curation, and issue triage.
Parameters—
Domainrag-retrievalagent-memory
Open SourceNo
WebsiteVisit
Luna-EFMhallucination-detectionguardrailsChainPollobservability
Capability Profile
Benchmark Scores
6 of 14 benchmarksLong-Context Retrieval1/5
Multi-Turn Recall1/2
MemoryBank
no dataCross-Session Memory1/1
Multi-Hop QA2/3
Agent Task Memory1/1
Personalization0/1
PerLTQA
no dataFactuality / Grounding0/1
RAGAS
no dataSources:Galileo AI vendor documentation; evaluated on HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering (Stanford / CMU, 1809)Galileo AI vendor documentation; evaluated on MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries (HKUST, 2401)Galileo AI vendor documentation; evaluated on AgentBench Memory Track (Tsinghua KEG, 2308)Galileo AI vendor documentation; evaluated on LoCoMo: Long-Term Conversational Memory Benchmark (Snap Research, 2402)Galileo AI vendor documentation; evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)Galileo AI vendor documentation; evaluated on LongMemEval: Benchmarking Chat Assistants on Long-Term Interactive Memory (Salesforce AI Research, 2410)