V

Voice Safety Classifier

Developed by Roblox
A voice content safety detection model based on WavLM base plus architecture, used to identify toxic content in voice chats
Downloads 11.55k
Release Time : 6/28/2024

Model Overview

This model is a large-scale classification model specifically designed to detect toxic content in voice chat audio, including profanity, explicit content, racial discrimination, bullying, and other types of violations.

Model Features

Multi-label classification
Capable of detecting multiple types of violations in voice content simultaneously, including profanity, explicit content, racial discrimination, and bullying.
High accuracy
Achieves an average accuracy of 94.48% on manually annotated datasets.
Large-scale training data
Fine-tuned using 2,374 hours of voice chat audio clips.

Model Capabilities

Voice content classification
Toxic content detection
Multi-label prediction

Use Cases

Content safety
Voice chat monitoring
Real-time monitoring of inappropriate content in voice chat platforms
Effectively identifies various types of inappropriate voice content.
Community management
Automatically flags potentially harmful voice content for manual review
Reduces manual review workload and improves moderation efficiency.
Featured Recommended AI Models
ยฉ 2025AIbase