A

Audio Source Separation

Developed by Awais
Audio source separation model trained on the Asteroid framework, optimized for speech separation tasks at 8kHz sampling rate
Downloads 30
Release Time : 4/2/2022

Model Overview

This model uses the ConvTasNet architecture and is trained for clean speech separation tasks on the Libri2Mix dataset, capable of separating different speaker voices from mixed audio

Model Features

Efficient separation
Adopts ConvTasNet architecture to achieve efficient speech separation at 8kHz sampling rate
Optimized training
Specially optimized based on the Libri2Mix dataset, suitable for clean speech separation scenarios
Lightweight
Moderate model parameter size, suitable for practical deployment applications

Model Capabilities

Dual-speaker speech separation
8kHz audio processing
Real-time audio source separation

Use Cases

Speech processing
Meeting recording enhancement
Separate different speaker voices in meeting recordings
SI-SDR improvement of 14.76dB
Speech recognition preprocessing
Provide cleaner single-speaker audio input for ASR systems
STOI improvement of 0.93
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase