B

Bigbird Pegasus Large Arxiv

Developed by google
BigBird is a Transformer model based on sparse attention, capable of handling longer sequences, suitable for tasks like long document summarization.
Downloads 8,528
Release Time : 3/2/2022

Model Overview

BigBirdPegasus is a Transformer model based on sparse attention, extending the capabilities of traditional Transformers, particularly suited for long-sequence tasks such as scientific paper summarization.

Model Features

Sparse Attention Mechanism
Employs block sparse attention mechanism, efficiently processing sequences up to 4096 in length with lower computational cost than traditional attention mechanisms.
Long Sequence Processing Capability
Particularly suitable for ultra-long sequence tasks, such as long document summarization and long-context question answering.
Multi-Task Adaptation
Performs well on various summarization tasks, including scientific papers and news summarization.

Model Capabilities

Scientific Paper Summarization
Long Document Summarization
Long-Context Question Answering

Use Cases

Academic Research
Scientific Paper Summarization
Generates concise and accurate summaries for scientific papers.
Achieves a ROUGE-1 score of 43.4702 on the arxiv dataset.
News Media
News Summarization
Generates summaries for news articles.
Achieves a ROUGE-1 score of 9.0885 on the cnn_dailymail dataset.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase