AI Tools.

Search

feature extraction

bge-large-en-v1.5

BGE-Large-EN-v1.5 is BAAI's highest-capacity English embedding model in the v1.5 series, producing 1024-dimensional vectors. It achieves top MTEB retrieval scores among its generation of English-only embedding models, at the cost of higher compute and storage than BGE-small or BGE-base. MIT licensed with ONNX export support.

Last reviewed

Use cases

  • High-precision semantic search where embedding quality is the primary constraint
  • Embedding for legal, medical, or technical domain retrieval requiring fine-grained distinction
  • MTEB benchmark baseline as a strong English embedding reference point
  • Re-ranking large candidate sets using embedding similarity
  • Knowledge base retrieval where 768-dim models underperform

Pros

  • Strong MTEB retrieval accuracy at 1024 dimensions
  • MIT license for commercial use
  • ONNX and text-embeddings-inference compatible for production deployment
  • Part of the well-maintained BAAI BGE family with documented benchmarks

Cons

  • 1024-dim output doubles storage cost vs. 512-dim alternatives
  • Higher inference compute than BGE-small or BGE-base
  • English-only; no multilingual or cross-lingual capability
  • May provide marginal gains over BGE-base for many standard retrieval tasks
  • Newer instruction-following embedding models are competitive at smaller sizes

FAQ

What is bge-large-en-v1.5 used for?

High-precision semantic search where embedding quality is the primary constraint. Embedding for legal, medical, or technical domain retrieval requiring fine-grained distinction. MTEB benchmark baseline as a strong English embedding reference point. Re-ranking large candidate sets using embedding similarity. Knowledge base retrieval where 768-dim models underperform.

Is bge-large-en-v1.5 free to use?

bge-large-en-v1.5 is an open-source model published on HuggingFace. License terms vary by model — check the model card for the specific license.

How do I run bge-large-en-v1.5 locally?

Most HuggingFace models can be loaded with transformers or the appropriate framework library. See the model card for framework-specific instructions and hardware requirements.

Tags

sentence-transformerspytorchonnxsafetensorsbertfeature-extractionsentence-similaritytransformersmtebenarxiv:2401.03462arxiv:2312.15503arxiv:2311.13534arxiv:2310.07554arxiv:2309.07597license:mitmodel-indexeval-resultstext-embeddings-inferenceendpoints_compatible