C

Chinese Macbert Base

Developed by hfl
MacBERT is an improved BERT model that uses a novel MLM as a corrective masked language model pre-training task, alleviating the discrepancy between pre-training and fine-tuning stages.
Downloads 22.48k
Release Time : 3/2/2022

Model Overview

MacBERT is an enhanced Chinese BERT model that uses similar words for masking instead of the traditional [MASK] token, reducing the gap between pre-training and fine-tuning phases.

Model Features

Corrective MLM
Uses similar words for masking instead of the traditional [MASK] token, reducing the gap between pre-training and fine-tuning stages.
Whole Word Masking
Employs whole word masking technology to enhance the model's understanding of complete vocabulary.
N-gram Masking
Supports N-gram masking to strengthen the model's ability to process continuous vocabulary.
Sentence Order Prediction
Integrates Sentence Order Prediction (SOP) tasks to improve the model's understanding of sentence structure.

Model Capabilities

Text Understanding
Text Generation
Text Classification
Named Entity Recognition
Question Answering System

Use Cases

Natural Language Processing
Chinese Text Classification
Used for tasks such as sentiment analysis and topic classification in Chinese text.
Named Entity Recognition
Identifies entities such as person names, place names, and organization names in Chinese text.
Question Answering System
Builds a Chinese question answering system to respond to user queries.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase