D

Distil Medium.en

Developed by distil-whisper
Distil-Whisper is a distilled version of the Whisper model, 6 times faster than the original, with a 49% reduction in size, while maintaining performance close to the original in English speech recognition tasks.
Downloads 186.85k
Release Time : 10/24/2023

Model Overview

Distil-Whisper is a lightweight variant obtained from the Whisper model through large-scale pseudo-label knowledge distillation, specifically designed for efficient English speech recognition.

Model Features

Efficient Inference
6 times faster than the original Whisper, suitable for real-time applications
Compact Model
49% reduction in size, reducing memory and storage requirements
High Performance
Performance remains within 1% word error rate (WER) of the original on out-of-distribution evaluation sets
Long-form Transcription Optimization
Uses a chunking algorithm, making long-form transcription 9 times faster than sequential algorithms

Model Capabilities

English Speech Recognition
Short-form audio transcription (<30 seconds)
Long-form audio transcription (>30 seconds)
Speculative decoding (as an auxiliary model to Whisper)

Use Cases

Speech Transcription
Meeting Minutes
Automatically transcribe English meeting content
Highly accurate text records
Podcast Transcription
Convert English podcast content into text
Quickly generate searchable text content
Assistive Technology
Real-time Caption Generation
Generate real-time captions for English video content
Low-latency caption display
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase