BAAI
BGE Large EN v1.5
High quality English embedding model. Best accuracy for English search.
About This Model
BGE Large EN v1.5, developed by BAAI, is a 335 million parameter BERT-based model designed for efficient feature extraction and embedding generation. This model excels in creating high-quality embeddings for text data, making it particularly useful for tasks such as semantic similarity, clustering, and information retrieval. Its ability to generate rich, context-aware embeddings within a 512-token context length ensures that it can handle a wide range of input sizes effectively.
Despite its relatively modest size, BGE Large EN v1.5 punches well above its weight in terms of performance and efficiency. It offers a balance between computational requirements and output quality, making it a strong choice for developers and researchers who need robust embeddings without the overhead of larger models. The model’s efficiency is further enhanced by its availability in quantized formats (Q8_0, FP16), which significantly reduce memory usage and improve inference speed, requiring only 0.8–1.1 GB of VRAM. This makes it suitable for deployment on a variety of hardware, including laptops and mid-range GPUs, without compromising on performance. Ideal users include those working on projects that require fast and accurate text embeddings, such as natural language processing applications, content recommendation systems, and data analysis tasks.
Check Your Hardware
See which quantizations of BGE Large EN v1.5 your hardware can run.
Quantization Options
| Quantization | Bits | File Size | VRAM Needed | RAM Needed | Quality |
|---|---|---|---|---|---|
| Q8_0 | 8 | 0.334 GB | 0.83 GB | 1.33 GB | 98% |
| FP16 | 16 | 0.624 GB | 1.12 GB | 1.62 GB | 100% |
Frequently Asked Questions
How much VRAM do I need to run BGE Large EN v1.5?
BGE Large EN v1.5 requires 0.83GB VRAM minimum with Q8_0 quantization. For full precision, you need 1.12GB VRAM.
What is the best quantization for BGE Large EN v1.5?
Q4_K_M offers the best balance of quality and VRAM usage. Q8_0 is near-lossless if you have enough VRAM.