HuggingFace

Distil-Whisper Large v3

Distilled Whisper. 6x faster than large-v3 with 1% accuracy loss.

0.76B parameterswhispermit1.92GB - 1.92GB VRAM

About This Model

Distil-Whisper Large v3, developed by HuggingFace, is an efficient and compact version of the popular Whisper architecture, designed for automatic speech recognition (ASR). With 0.76 billion parameters, this model offers a compelling balance between performance and resource requirements, making it suitable for a wide range of applications, from real-time transcription to voice-controlled interfaces. It excels in recognizing and transcribing spoken words with high accuracy, even in noisy environments, thanks to its robust training on diverse datasets.

In its size class, Distil-Whisper Large v3 punches well above its weight. Despite being significantly smaller than some of its full-sized counterparts, it maintains a high level of accuracy and efficiency. This makes it particularly appealing for users who need powerful ASR capabilities but have limited computational resources. The model’s quantization options, such as Q8_0, further enhance its efficiency, requiring only 1.9 GB of VRAM, which is manageable even on mid-range GPUs and some high-end CPUs.

This model is ideal for developers and hobbyists looking to integrate ASR into their projects without the overhead of cloud services. Realistic hardware for running Distil-Whisper Large v3 includes modern laptops with dedicated GPUs, high-end desktops, and edge devices with sufficient RAM and processing power. Its low VRAM requirement and efficient quantization make it accessible to a broader audience, ensuring that it can be deployed in a variety of settings, from personal projects to small-scale commercial applications.

Check Your Hardware

See which quantizations of Distil-Whisper Large v3 your hardware can run.

Quantization Options

QuantizationBitsFile SizeVRAM NeededRAM NeededQuality
Q8_081.415 GB1.92 GB2.42 GB
96%

Frequently Asked Questions

How much VRAM do I need to run Distil-Whisper Large v3?

Distil-Whisper Large v3 requires 1.92GB VRAM minimum with Q8_0 quantization. For full precision, you need 1.92GB VRAM.

What is the best quantization for Distil-Whisper Large v3?

Q4_K_M offers the best balance of quality and VRAM usage. Q8_0 is near-lossless if you have enough VRAM.