L

Long T5 Tglobal Base

Developed by google
LongT5 is a text-to-text transformation model based on the T5 architecture, employing transient global attention mechanism for efficient processing of long sequence inputs
Downloads 71.38k
Release Time : 4/16/2022

Model Overview

LongT5 is a Transformer model based on encoder-decoder architecture, efficiently handling long sequences (up to 16,384 tokens) through local attention or transient-global attention mechanisms, particularly suitable for generation tasks requiring long text processing

Model Features

Long Sequence Processing Capability
Supports long sequence inputs up to 16,384 tokens, achieving efficient processing through sparse attention mechanisms
Transient Global Attention
Adopts innovative transient-global attention mechanism, reducing computational complexity while maintaining performance
Generative Pre-training
Uses Pegasus-like generative denoising pre-training method to optimize text generation capabilities

Model Capabilities

Long text summarization
Long document question answering systems
Text-to-text transformation

Use Cases

Text Summarization
Automatic Summarization of Long Documents
Generates concise summaries for long documents such as research papers and legal documents
Excels in long text summarization tasks
Question Answering Systems
Long Document Question Answering
Extracts information from long documents to answer complex questions
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase