D

Distilbert Punctuator Zh

Developed by Qishuai
A Chinese punctuation restoration model fine-tuned based on DistilBertForTokenClassification, specifically designed to add punctuation marks to unpunctuated Chinese text
Downloads 103
Release Time : 3/2/2022

Model Overview

This model is fine-tuned on the distilled version of bert-base-chinese and is used to automatically add punctuation marks to unpunctuated Simplified Chinese text

Model Features

Lightweight model
Based on the distilled version of DistilBert, it reduces the model size while maintaining performance
Multi-punctuation support
Supports the restoration of various Chinese punctuation marks such as commas, enumeration commas, exclamation marks, periods, and question marks
News corpus training
Trained using 2014 People's Daily news corpus, suitable for formal text scenarios

Model Capabilities

Chinese punctuation restoration
Automatic text punctuation
Unpunctuated text processing

Use Cases

Text processing
Speech-to-text post-processing
Adding punctuation marks to unpunctuated text output by speech recognition systems
Improves text readability
Ancient text digitization
Adding modern punctuation to unpunctuated text generated during the digitization of ancient literature
Facilitates reading for modern audiences
News processing
News manuscript processing
Automatically adding punctuation to unpunctuated news manuscripts quickly entered by reporters
Enhances editing efficiency
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase