E

Erlangshen DeBERTa V2 97M Chinese

Developed by IDEA-CCNL
A Chinese DeBERTa-v2 base model specialized in natural language understanding tasks, employing Whole Word Masking with 97 million parameters.
Downloads 178
Release Time : 7/19/2022

Model Overview

An enhanced BERT model based on disentangled attention mechanism, designed specifically for Chinese natural language understanding tasks, suitable for text classification, sentiment analysis, and similar scenarios.

Model Features

Whole Word Masking
Utilizes Whole Word Masking during pre-training to enhance the model's understanding of Chinese vocabulary.
Disentangled Attention Mechanism
Based on DeBERTa-v2 architecture, employs disentangled attention mechanism to improve model performance.
Chinese Optimization
Specially optimized for Chinese language characteristics, trained on 180G Wudao corpus.

Model Capabilities

Text classification
Sentiment analysis
Natural language inference
Masked language modeling

Use Cases

Text Analysis
News Classification
Classify news texts
57.1% accuracy on TNEWS dataset
App Classification
Classify application descriptions
59.77% accuracy on IFLYTEK dataset
Semantic Understanding
Natural Language Inference
Determine logical relationships between sentences
75.68% accuracy on OCNLI dataset
Chinese Inference
Perform Chinese language inference tasks
80.7% accuracy on CMNLI dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase