Erlangshen DeBERTa V2 320M Chinese
Chinese pre-trained language model based on DeBERTa-v2 architecture with 320 million parameters, excelling in natural language understanding tasks
Downloads 186
Release Time : 9/21/2022
Model Overview
Chinese DeBERTa-v2-Large model employing Whole Word Masking technique, trained on WuDao corpus, suitable for various Chinese NLP tasks
Model Features
Whole Word Masking
Employs whole word masking strategy during pre-training to enhance the model's understanding of Chinese words
Disentangled Attention Mechanism
Based on DeBERTa-v2 architecture, uses disentangled attention mechanism to improve model performance
Large-scale Pre-training
Pre-trained on 180G version of WuDao corpus, possessing strong language understanding capabilities
Model Capabilities
Text Completion
Sentiment Analysis
Text Classification
Natural Language Inference
Masked Language Modeling
Use Cases
Text Analysis
Sentiment Analysis
Analyze text sentiment orientation
Achieved 74.98% accuracy on AFQMC dataset
News Classification
Classify news texts
Achieved 58.17% accuracy on TNEWS dataset
Natural Language Understanding
Natural Language Inference
Determine logical relationship between two texts
Achieved 80.22% accuracy on OCNLI dataset
Text Completion
Predict masked text content
Can accurately predict masked words like 'Li' River in examples
Featured Recommended AI Models
Š 2025AIbase