Bloomz 3b Guardrail
B

Bloomz 3b Guardrail

Developed by cmarkea
Bloomz-3b-guardrail is a text classification model fine-tuned based on Bloomz-3b-sft-chat, used to detect text toxicity in five modes.
Downloads 249
Release Time : 12/1/2023

Model Overview

This model aims to monitor and control the output of generative models, and detect the toxicity level of text in five modes: obscene content, explicit pornographic content, identity attacks, insults, and threats.

Model Features

Multimodal toxicity detection
It can detect the toxicity of text in five modes: obscene content, explicit pornographic content, identity attacks, insults, and threats.
High correlation
The model output is highly correlated with the judge's score, with a Pearson correlation of approximately 80.
Multilingual support
Supports toxicity detection in English and French.

Model Capabilities

Text toxicity detection
Multimodal classification
Multilingual processing

Use Cases

Content moderation
Social media content monitoring
Used to detect harmful content on social media, such as insults and threats.
It can accurately identify multiple toxicity modes and help platforms handle违规 content in a timely manner.
Output control of generative models
Monitor the output of generative models to ensure that they do not produce harmful content.
Effectively reduce the toxicity of generated content and improve the user experience.
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase