P

Prompt Injection Defender Large V0

Developed by testsavantai
The TestSavantAI model is a set of classifiers specifically designed to defend against prompt injection and jailbreak attacks in large language models (LLMs). The tiny version is based on the BERT-tiny architecture, balancing security and computational efficiency.
Downloads 23
Release Time : 11/27/2024

Model Overview

This model is used to detect and intercept malicious prompt injections and jailbreak attempts targeting AI systems, protecting language models from misuse.

Model Features

Guard Effectiveness Score (GES)
An innovative composite metric that combines Attack Success Rate (ASR) and False Rejection Rate (FRR) to evaluate model robustness.
Multi-size Variants
Offers models of different sizes from tiny to large to meet varying performance and computational efficiency requirements.
ONNX Support
Provides ONNX runtime versions to optimize inference performance.

Model Capabilities

Malicious Prompt Detection
Jailbreak Attack Interception
Text Classification
AI Security Protection

Use Cases

AI Security
ChatGPT Protection
Detects and intercepts jailbreak prompts targeting ChatGPT.
Effectively reduces the success rate of malicious prompt injections.
Enterprise AI System Protection
Protects enterprise-deployed AI systems from prompt injection attacks.
Reduces the risk of system misuse.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase