Erlangshen DeBERTa V2 710M Chinese
This is a 710M parameter DeBERTa-v2 model focused on Chinese natural language understanding tasks. It is pre-trained using the whole-word masking method, providing strong support for the Chinese NLP field.
Downloads 246
Release Time : 8/16/2022
Model Overview
The Erlangshen-DeBERTa-v2-710M-Chinese model is a Chinese pre-trained model based on the DeBERTa-v2 architecture. It is good at handling natural language understanding tasks and uses the whole-word masking method to improve the pre-training effect.
Model Features
Whole-word masking pre-training
Adopt the whole-word masking (wwm) method to improve the pre-training effect
Powerful language understanding ability
Based on the DeBERTa-v2-XLarge architecture with 710 million parameters, it has strong language understanding ability
Chinese optimization
Specifically optimized for Chinese NLP tasks and performs excellently in multiple Chinese NLU tasks
Model Capabilities
Text understanding
Semantic analysis
Text completion
Use Cases
Natural language understanding
Sentiment analysis
Analyze the emotional tendency in the text
Text classification
Classify the text content
Language model tasks
Masked language modeling
Predict the masked words
Performs better than RoBERTa-base/large in multiple Chinese NLU tasks
Featured Recommended AI Models
Š 2025AIbase