D

Distil Small.en

Developed by distil-whisper
Distil-Whisper is a distilled version of the Whisper model, 6x faster with 49% smaller size, achieving near 1% WER on out-of-distribution evaluation sets.
Downloads 33.51k
Release Time : 12/6/2023

Model Overview

Distil-Whisper is an automatic speech recognition (ASR) model specifically designed for English speech recognition. It is a distilled version of the Whisper model, optimized for inference speed and model size while maintaining high recognition accuracy.

Model Features

Efficient Inference
6x faster than the original Whisper model, suitable for real-time applications
Compact Model
49% smaller size, ideal for memory-constrained on-device applications
High Performance
Achieves near 1% WER on out-of-distribution evaluation sets
Long-form Transcription Optimization
Uses chunking algorithm, 9x faster than sequential processing

Model Capabilities

English Speech Recognition
Short-form audio transcription (<30 seconds)
Long-form audio transcription (>30 seconds)
Speculative decoding (as an auxiliary model to Whisper)

Use Cases

Speech Transcription
Meeting Minutes
Convert meeting recordings into text transcripts
Fast and accurate transcription results
Podcast Transcription
Convert podcast content into text
Efficient transcription for long-form audio
On-device Applications
Mobile Voice Assistants
Implement speech recognition on mobile devices
Small model size suitable for on-device deployment
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase