E

Erlangshen DeBERTa V2 710M Chinese

Developed by IDEA-CCNL
This is a 710M parameter DeBERTa-v2 model focused on Chinese natural language understanding tasks. It is pre-trained using the whole-word masking method, providing strong support for the Chinese NLP field.
Downloads 246
Release Time : 8/16/2022

Model Overview

The Erlangshen-DeBERTa-v2-710M-Chinese model is a Chinese pre-trained model based on the DeBERTa-v2 architecture. It is good at handling natural language understanding tasks and uses the whole-word masking method to improve the pre-training effect.

Model Features

Whole-word masking pre-training
Adopt the whole-word masking (wwm) method to improve the pre-training effect
Powerful language understanding ability
Based on the DeBERTa-v2-XLarge architecture with 710 million parameters, it has strong language understanding ability
Chinese optimization
Specifically optimized for Chinese NLP tasks and performs excellently in multiple Chinese NLU tasks

Model Capabilities

Text understanding
Semantic analysis
Text completion

Use Cases

Natural language understanding
Sentiment analysis
Analyze the emotional tendency in the text
Text classification
Classify the text content
Language model tasks
Masked language modeling
Predict the masked words
Performs better than RoBERTa-base/large in multiple Chinese NLU tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase