Randeng Pegasus 523M Chinese
A Chinese version of the PAGASUS-large model specialized in text summarization tasks, trained on the PEGASUS architecture with optimizations for Chinese tokenization.
Downloads 329
Release Time : 6/9/2022
Model Overview
This model is a Chinese text summarization model based on the PEGASUS architecture, pre-trained on 180G of WuDao corpus, and integrates Jieba tokenizer with BERT tokenizer to address Chinese tokenization stability issues.
Model Features
Chinese Optimization
Innovatively combines Jieba tokenizer and BERT tokenizer to address Chinese tokenization stability issues.
Large-Scale Pre-training
Uses the 180G WuDao corpus as pre-training data to enhance model performance.
Multi-Version Support
Provides both a base model and a fine-tuned version for Chinese summarization datasets to meet diverse needs.
Model Capabilities
Text Summarization Generation
Chinese Text Processing
Use Cases
News Summarization
News Content Summarization
Generates concise summaries of lengthy news articles, extracting key information.
Produces brief summaries that capture the core content of the original text.
Document Processing
Long Document Summarization
Generates summaries for technical documents, reports, and other lengthy texts.
Extracts the core content of documents and produces concise summaries.
Featured Recommended AI Models