P

Pegasus Xsum

Developed by google
PEGASUS is a Transformer-based pretrained model specifically designed for abstractive text summarization tasks.
Downloads 144.72k
Release Time : 3/2/2022

Model Overview

PEGASUS is a Transformer architecture-based pretrained model, specifically designed for abstractive text summarization tasks. It learns to generate high-quality summaries by pretraining on large-scale text data.

Model Features

Hybrid and Random Training
Trained simultaneously on C4 and HugeNews datasets with sample count-weighted mixing ratios and random sampling of important sentences.
Dynamic Sentence Gap Ratio
Uniformly samples sentence gap ratios between 15% to 45% during training to enhance model adaptability.
Importance Score Noise
Adds 20% uniform noise to importance scores during sentence sampling to improve model robustness.
Improved Tokenizer
Updated SentencePiece tokenizer to support encoding newline characters, preserving paragraph segmentation information.

Model Capabilities

Text Summarization Generation
Multi-document Summarization
Abstractive Summarization

Use Cases

News Summarization
CNN/DailyMail News Summarization
Generates concise summaries for CNN/DailyMail news articles
ROUGE-1/2/L: 44.16/21.56/41.30
Academic Paper Summarization
arXiv Paper Summarization
Generates summaries for arXiv academic papers
ROUGE-1/2/L: 44.21/16.95/25.67
Legal Document Summarization
BigPatent Patent Summarization
Generates summaries for patent documents
ROUGE-1/2/L: 52.29/33.08/41.66
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase