Pegasus Cnn Dailymail
PEGASUS is an abstractive summarization pretraining model based on gap sentence generation, focusing on text summarization tasks, with performance enhanced through mixed dataset training and various optimization strategies.
Downloads 37.32k
Release Time : 3/2/2022
Model Overview
PEGASUS is a pretrained model specifically designed for text summarization tasks. It achieves pretraining by extracting gap sentences and demonstrates excellent performance across multiple datasets.
Model Features
Mixed Dataset Training
Trained simultaneously on C4 and HugeNews datasets with sample-size weighted mixing to enhance model generalization.
Random Sampling Strategy
Adopts 15%-45% uniform sampling for gap sentence ratio with 20% uniform noise added to importance scores to improve model robustness.
Tokenizer Optimization
Upgraded sentencepiece tokenizer to support newline character encoding, enhancing processing capability for specific text formats.
Extended Training
Trained for 1.5 million steps (originally 500k steps), observing slower pretraining perplexity convergence but ultimately improved performance.
Model Capabilities
Text Summarization Generation
Multi-dataset Adaptation
Abstractive Summarization
Use Cases
News Summarization
CNN/DailyMail Summarization
Generates concise summaries of news articles
ROUGE-1/2/L: 44.16/21.56/41.30
XSum Summarization
Generates extreme summarization (single-sentence summaries)
ROUGE-1/2/L: 47.60/24.83/39.64
Academic Paper Summarization
arXiv Paper Summarization
Generates summaries of academic papers
ROUGE-1/2/L: 44.21/16.95/25.67
PubMed Summarization
Generates summaries of medical literature
ROUGE-1/2/L: 45.97/20.15/28.25
Technical Document Summarization
BigPatent Summarization
Generates summaries of patent documents
ROUGE-1/2/L: 52.29/33.08/41.66
WikiHow Summarization
Generates summaries of how-to guides
ROUGE-1/2/L: 46.39/22.12/38.41
Featured Recommended AI Models
Š 2025AIbase