Sew Tiny 100k
SEW-tiny is a compressed and efficient speech pretraining model developed by ASAPP Research, pretrained on 16kHz sampled speech audio, suitable for various downstream speech tasks.
Downloads 1,080
Release Time : 3/2/2022
Model Overview
SEW-tiny is an efficient speech pretraining model designed for tasks like automatic speech recognition, improving inference speed while maintaining performance through optimized architecture.
Model Features
Efficient Inference
Achieves 1.9x faster inference compared to wav2vec 2.0
Performance Optimization
Reduces word error rate by 25-50% under similar inference time conditions
Compressed Architecture
Specially designed compressed and efficient architecture, optimizing performance-efficiency trade-off
Model Capabilities
Speech Recognition
Speaker Recognition
Intent Classification
Emotion Recognition
Use Cases
Speech Processing
Automatic Speech Transcription
Convert speech content into text
13.5% reduction in word error rate on the LibriSpeech dataset
Voice Assistants
Speech recognition module for intelligent voice assistants
Featured Recommended AI Models