Jina AI

Jina Reranker Tiny EN

Tiny English reranker. Only 67MB. Use with embedding models for better search.

0.033B parametersbertapache-2.08K context0.15GB - 0.15GB VRAM

About This Model

The Jina Reranker Tiny EN is a compact BERT-based model designed for re-ranking text results, enhancing the relevance of search outputs by refining the order of documents or snippets. With just 0.033 billion parameters, this model is exceptionally lightweight, making it highly efficient for devices with limited computational resources. It excels in scenarios where quick, on-the-fly re-ranking is needed without the overhead of larger, more resource-intensive models. The model supports a context length of 8192 tokens, which is quite generous for its size, allowing it to handle longer texts effectively.

Despite its small size, the Jina Reranker Tiny EN performs well, often delivering results that are comparable to larger models in terms of accuracy and relevance. This makes it an excellent choice for applications where efficiency and speed are critical, such as mobile devices, edge computing, or any environment with strict memory constraints. The model’s low VRAM requirement of 0.1 GB means it can run smoothly on a wide range of hardware, from older laptops to modern smartphones. Developers and researchers looking to integrate a lightweight yet powerful re-ranking solution into their projects will find this model particularly useful.

Check Your Hardware

See which quantizations of Jina Reranker Tiny EN your hardware can run.

Quantization Options

QuantizationBitsFile SizeVRAM NeededRAM NeededQuality
FP16160.067 GB0.15 GB0.3 GB
85%

Frequently Asked Questions

How much VRAM do I need to run Jina Reranker Tiny EN?

Jina Reranker Tiny EN requires 0.15GB VRAM minimum with FP16 quantization. For full precision, you need 0.15GB VRAM.

What is the best quantization for Jina Reranker Tiny EN?

Q4_K_M offers the best balance of quality and VRAM usage. Q8_0 is near-lossless if you have enough VRAM.