D

Distil Large V3.5

Developed by distil-whisper
Distil-Whisper is a knowledge-distilled version of OpenAI Whisper-Large-v3, achieving efficient speech recognition through large-scale pseudo-label training.
Downloads 4,804
Release Time : 12/5/2024

Model Overview

Distil-Large-v3.5 is the latest addition to the Distil-Whisper English series, offering improved performance while maintaining efficiency. Trained on 98,000 hours of data using 'patient' teacher model strategy and SpecAugment data augmentation techniques.

Model Features

Efficient Inference
Approximately 1.5x faster than Whisper-Large-v3-Turbo while maintaining comparable accuracy
Knowledge Distillation Optimization
Utilizes 'patient' teacher model strategy and SpecAugment data augmentation, trained on 98,000 hours of data
Speculative Decoding Compatibility
Suitable as a draft model for Whisper-Large-v3 speculative decoding, enabling ~2x inference acceleration

Model Capabilities

Short-form speech transcription
Long-form speech transcription
Timestamp generation
English speech recognition

Use Cases

Speech Transcription
Meeting Minutes
Convert meeting recordings into text transcripts
Word Error Rate (WER) ~7.08%
Podcast Transcription
Convert long-form audio content into text
Long-form WER ~11.39%
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase