Pangolin Guard Large
A lightweight model based on ModernBERT (large model edition), specifically designed to identify malicious prompts (i.e., prompt injection attacks).
Downloads 72
Release Time : 3/11/2025
Model Overview
Pangolin Guard is a lightweight model designed to identify and defend against security challenges such as prompt injection and jailbreaking in large language model (LLM) applications. It effectively prevents sensitive data leaks or deviations in model behavior.
Model Features
Lightweight Design
The model is designed to be lightweight, suitable for self-hosting and low-cost deployment.
High Accuracy
Demonstrates high accuracy and F1 scores in specialized benchmark tests, effectively identifying malicious prompts.
Open Source
Fully open-source, facilitating community use and improvement.
Low Over-Defense Tendency
Passes the NotInject test, measuring the over-defense tendency of protective models to ensure benign inputs are not misjudged.
Model Capabilities
Identify malicious prompts
Defend against prompt injection attacks
Detect jailbreaking attacks
Protect sensitive data
Use Cases
AI Agents and Chat Interfaces
Self-Hosted Defense Mechanism
Adds a self-hosted, low-cost prompt injection attack defense mechanism for AI agents and chat interfaces.
Effectively prevents sensitive data leaks and deviations in model behavior.
Security Protection
Privacy Violation Attempt Detection
Evaluates privacy violation attempts and boundary-testing queries posed via indirect prompt injection attacks.
High accuracy in identifying malicious behavior.
Featured Recommended AI Models
Š 2025AIbase