B

Bigbird Pegasus Large Bigpatent

Developed by google
BigBird is a Transformer model based on sparse attention, capable of processing sequences up to 4096 in length, suitable for tasks like long document summarization.
Downloads 945
Release Time : 3/2/2022

Model Overview

BigBird employs block sparse attention mechanisms to replace conventional attention, enabling efficient processing of long sequences with lower computational costs, excelling in tasks like long document summarization.

Model Features

Block Sparse Attention Mechanism
Uses block sparse attention instead of conventional attention, significantly reducing computational costs for long sequence processing.
Long Sequence Processing Capability
Efficiently handles sequences up to 4096 in length, making it suitable for long document tasks.
Flexible Configuration
Supports adjusting block size and random block count to balance performance and computational resources.

Model Capabilities

Long Text Summarization
Long Context Understanding

Use Cases

Document Processing
Patent Document Summarization
Generates concise summaries for lengthy patent documents
Fine-tuned on the big_patent dataset
Long Document Question Answering
Answers questions based on long document content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase