R

Randeng Pegasus 523M Chinese

Developed by IDEA-CCNL
A Chinese version of the PAGASUS-large model specialized in text summarization tasks, trained on the PEGASUS architecture with optimizations for Chinese tokenization.
Downloads 329
Release Time : 6/9/2022

Model Overview

This model is a Chinese text summarization model based on the PEGASUS architecture, pre-trained on 180G of WuDao corpus, and integrates Jieba tokenizer with BERT tokenizer to address Chinese tokenization stability issues.

Model Features

Chinese Optimization
Innovatively combines Jieba tokenizer and BERT tokenizer to address Chinese tokenization stability issues.
Large-Scale Pre-training
Uses the 180G WuDao corpus as pre-training data to enhance model performance.
Multi-Version Support
Provides both a base model and a fine-tuned version for Chinese summarization datasets to meet diverse needs.

Model Capabilities

Text Summarization Generation
Chinese Text Processing

Use Cases

News Summarization
News Content Summarization
Generates concise summaries of lengthy news articles, extracting key information.
Produces brief summaries that capture the core content of the original text.
Document Processing
Long Document Summarization
Generates summaries for technical documents, reports, and other lengthy texts.
Extracts the core content of documents and produces concise summaries.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase