M

Meta Llama Guard 2 8B

Developed by meta-llama
An 8-billion-parameter Llama 3-based security protection model for safety classification of LLM inputs and responses
Downloads 12.85k
Release Time : 4/17/2024

Model Overview

Llama Guard 2 is a security protection model based on Llama 3, capable of classifying content in LLM inputs (prompt classification) and LLM responses (response classification) to identify potentially harmful content.

Model Features

Safety Classification
Can perform safety classification on LLM inputs and responses to identify potentially harmful content
Multi-category Detection
Based on MLCommons hazard classification standards, can detect 11 different categories of harmful content
Probability Scoring
Generates classifier scores by analyzing the probability of the first token, allowing binary decisions through threshold setting

Model Capabilities

Text Classification
Content Safety Detection
Harmful Content Identification

Use Cases

Content Safety
LLM Input Filtering
Detects whether user inputs contain harmful content
Prevents harmful content from entering the LLM processing pipeline
LLM Response Filtering
Detects whether LLM-generated responses contain harmful content
Ensures LLM outputs comply with safety standards
Compliance Monitoring
Compliance Check
Ensures LLM interactions comply with laws, regulations, and platform policies
Reduces legal and reputational risks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase