Back to Arena
Cohere Embed
by Cohere Inc.
System Card
OrganizationCohere Inc.
Released2021-01
Architectureexternal-memory-network / multilingual multimodal embedding API
DetailsCohere Embed v3 (released November 2023) is a multimodal embedding model supporting 100+ languages and image+text encoding, achieving SOTA on MTEB at release. The model uses compression-aware training to produce high-quality int8 quantized embeddings. Embed 4 (2025) added further multimodal improvements at $0.12/MTok.
Parameters—
Domainrag-retrieval
Open SourceNo
WebsiteVisit
multilingualmultimodalint8-compressionenterpriseMTEB
Capability Profile
Benchmark Scores
5 of 14 benchmarksMulti-Turn Recall0/2
LoCoMo
no dataMemoryBank
no dataCross-Session Memory0/1
LongMemEval
no dataMulti-Hop QA2/3
Agent Task Memory0/1
AgentBench-Mem
no dataPersonalization0/1
PerLTQA
no dataFactuality / Grounding1/1
Sources:Cohere Embed vendor documentation; evaluated on HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering (Stanford / CMU, 1809)Cohere Embed vendor documentation; evaluated on LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding (Tsinghua KEG, 2308)Cohere Embed vendor documentation; evaluated on MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries (HKUST, 2401)Cohere Embed vendor documentation; evaluated on RAGAS: Automated Evaluation of Retrieval-Augmented Generation (Exploding Gradients, 2309)Cohere Embed vendor documentation; evaluated on RULER: What's the Real Context Size of Your Long-Context Language Models (NVIDIA, 2404)