C

Chinese Bert Wwm Ext

Developed by hfl
A Chinese pre-trained BERT model employing whole word masking strategy, aimed at accelerating Chinese natural language processing research.
Downloads 24.49k
Release Time : 3/2/2022

Model Overview

This model is a Chinese pre-trained model developed based on the official Google BERT project, utilizing whole word masking strategy, suitable for various Chinese natural language processing tasks.

Model Features

Whole Word Masking Strategy
Employs whole word masking strategy, which better captures semantic information of Chinese words compared to traditional masking strategies.
Chinese Optimization
Specifically optimized for Chinese language characteristics, suitable for Chinese natural language processing tasks.
Pre-trained Model
Provides pre-trained models that can be directly used for downstream tasks or fine-tuning.

Model Capabilities

Text Classification
Named Entity Recognition
Question Answering System
Text Generation

Use Cases

Natural Language Processing
Chinese Text Classification
Used for tasks such as sentiment analysis and topic classification of Chinese texts.
Named Entity Recognition
Identifies entities such as person names, place names, and organization names in Chinese texts.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase