Long T5 Tglobal Base
LongT5 is a text-to-text transformation model based on the T5 architecture, employing transient global attention mechanism for efficient processing of long sequence inputs
Downloads 71.38k
Release Time : 4/16/2022
Model Overview
LongT5 is a Transformer model based on encoder-decoder architecture, efficiently handling long sequences (up to 16,384 tokens) through local attention or transient-global attention mechanisms, particularly suitable for generation tasks requiring long text processing
Model Features
Long Sequence Processing Capability
Supports long sequence inputs up to 16,384 tokens, achieving efficient processing through sparse attention mechanisms
Transient Global Attention
Adopts innovative transient-global attention mechanism, reducing computational complexity while maintaining performance
Generative Pre-training
Uses Pegasus-like generative denoising pre-training method to optimize text generation capabilities
Model Capabilities
Long text summarization
Long document question answering systems
Text-to-text transformation
Use Cases
Text Summarization
Automatic Summarization of Long Documents
Generates concise summaries for long documents such as research papers and legal documents
Excels in long text summarization tasks
Question Answering Systems
Long Document Question Answering
Extracts information from long documents to answer complex questions
Featured Recommended AI Models
Š 2025AIbase