S

Shieldgemma 27b

Developed by google
ShieldGemma is a series of safety content moderation models based on Gemma 2, designed to moderate content across four hazard categories (sexual content, harmful content, hate speech, and harassment).
Downloads 65
Release Time : 7/16/2024

Model Overview

ShieldGemma is a decoder-only large language model supporting English, with open weights, used for safety content moderation.

Model Features

Multi-Hazard Category Moderation
Moderates content across four hazard categories (sexual content, harmful content, hate speech, and harassment).
Open Weights
Model weights are open, supporting custom deployment and usage.
High Performance
Outperforms similar open-source models in multiple benchmark tests.
Flexible Deployment
Supports single-GPU and multi-GPU deployment, offering various usage methods.

Model Capabilities

Text Classification
Content Safety Moderation
Generative AI Content Filtering

Use Cases

Content Moderation
User Input Filtering
Moderates user input content to ensure compliance with safety policies.
Identifies and filters user input that violates safety policies.
Model Output Filtering
Moderates AI-generated content to ensure compliance with safety policies.
Identifies and filters AI-generated content that violates safety policies.
Social Media
Hate Speech Detection
Detects hate speech content on social media.
Effectively identifies hate speech based on protected attributes such as race and gender.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase