M

MD Judge V0 2 Internlm2 7b

Developed by OpenSafetyLab
A safety protection tool fine-tuned based on internlm2-7b-chat, providing human-readable judgment explanations and fine-grained harm scoring
Downloads 1,823
Release Time : 7/21/2024

Model Overview

MD-Judge-v0.2-internlm2_7b is a large language model safety protection tool that provides reliable reference standards for safety evaluation, featuring chain-of-thought reasoning and fine-grained harm differentiation capabilities

Model Features

Chain-of-Thought Reasoning
Step-by-step reasoning to reach the final judgment, providing detailed reasoning process
Fine-Grained Harm Scoring
Precise evaluation based on a 5-level scoring standard (from completely harmless to extremely harmful)
Human-Readable Explanation
Provides clear judgment explanations and violation categories (if deemed unsafe)

Model Capabilities

Safety Evaluation
Harm Scoring
Violation Detection
Text Analysis

Use Cases

AI Safety
Dialogue Safety Evaluation
Evaluating potential harmful content in AI dialogues
Provides safe/unsafe judgment and specific scores
Content Moderation
Identifying and flagging harmful or dangerous content
Identifies and classifies violating content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase