C

Chinese Macbert Large

Developed by hfl
MacBERT is an improved Chinese BERT model that employs M as a corrective masked language model pre-training task, alleviating the inconsistency between pre-training and fine-tuning stages.
Downloads 13.05k
Release Time : 3/2/2022

Model Overview

MacBERT is an improved Chinese BERT model that enhances performance in Chinese natural language processing tasks by using similar words for masking instead of traditional [MASK] tokens, combined with techniques like whole-word masking, N-gram masking, and sentence order prediction.

Model Features

Corrective MLM
Uses similar words for masking instead of [MASK] tokens, alleviating inconsistency between pre-training and fine-tuning stages
Whole-word masking
Employs whole-word masking to enhance the model's understanding of Chinese words
N-gram masking
Supports N-gram level masking to improve the model's comprehension of long texts
Sentence order prediction
Incorporates sentence order prediction tasks to enhance the model's understanding of text coherence

Model Capabilities

Chinese text understanding
Text classification
Named entity recognition
Question answering systems
Text similarity calculation

Use Cases

Natural Language Processing
Chinese text classification
Used for tasks like sentiment analysis and topic classification in Chinese text
Named entity recognition
Identifies entities such as person names, locations, and organizations in Chinese text
Question answering systems
Builds Chinese question answering systems to respond to text-based questions
Featured Recommended AI Models
ยฉ 2025AIbase