U

Unbiased Toxic Roberta

Developed by unitary
A toxic comment classification tool based on PyTorch Lightning and Hugging Face Transformers, providing three models trained for Jigsaw competitions
Downloads 52.70k
Release Time : 3/2/2022

Model Overview

Detoxify is a tool for detecting text toxicity, containing three models trained for different Jigsaw competitions, capable of identifying various types of toxic content such as threats, obscenity, insults, and identity-based hate.

Model Features

Multi-Model Support
Provides three models trained for different Jigsaw competitions, suitable for toxicity detection in various scenarios
Multilingual Support
The multilingual model supports toxicity detection in 7 languages
Bias Minimization
The unbiased model is specifically optimized to reduce bias related to identity mentions

Model Capabilities

Toxic Content Detection
Multi-Label Classification
Multilingual Text Analysis
Bias Evaluation

Use Cases

Content Moderation
Social Media Comment Moderation
Automatically identify and flag toxic comments on social media platforms
Helps moderators quickly identify comments that need attention
Forum Content Filtering
Filter out insulting, threatening, or hateful speech in online forums
Maintains a healthy discussion environment
Academic Research
Language Bias Research
Study toxic expressions across different languages and cultural backgrounds
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase