Distilbert Punctuator Zh
D
Distilbert Punctuator Zh
Developed by Qishuai
A Chinese punctuation restoration model fine-tuned based on DistilBertForTokenClassification, specifically designed to add punctuation marks to unpunctuated Chinese text
Downloads 103
Release Time : 3/2/2022
Model Overview
This model is fine-tuned on the distilled version of bert-base-chinese and is used to automatically add punctuation marks to unpunctuated Simplified Chinese text
Model Features
Lightweight model
Based on the distilled version of DistilBert, it reduces the model size while maintaining performance
Multi-punctuation support
Supports the restoration of various Chinese punctuation marks such as commas, enumeration commas, exclamation marks, periods, and question marks
News corpus training
Trained using 2014 People's Daily news corpus, suitable for formal text scenarios
Model Capabilities
Chinese punctuation restoration
Automatic text punctuation
Unpunctuated text processing
Use Cases
Text processing
Speech-to-text post-processing
Adding punctuation marks to unpunctuated text output by speech recognition systems
Improves text readability
Ancient text digitization
Adding modern punctuation to unpunctuated text generated during the digitization of ancient literature
Facilitates reading for modern audiences
News processing
News manuscript processing
Automatically adding punctuation to unpunctuated news manuscripts quickly entered by reporters
Enhances editing efficiency
Featured Recommended AI Models