Long T5 Tglobal Xl
LongT5 is a Transformer-based text-to-text model specifically designed to handle long sequence inputs, supporting up to 16,384 tokens.
Downloads 336
Release Time : 6/14/2022
Model Overview
LongT5 is an extended version of the T5 model, adopting an encoder-decoder architecture. It efficiently processes long sequence texts through the transient global attention mechanism and is particularly suitable for tasks that require handling long inputs, such as summary generation and question answering.
Model Features
Transient global attention
Efficiently processes long sequences through a sparse attention pattern, supporting up to 16,384 tokens of input.
Text-to-text framework
Adopts a unified text-to-text conversion framework, supporting multiple generation tasks.
Pegasus-based pre-training
Uses a denoising generative pre-training method similar to Pegasus.
Model Capabilities
Long text summary generation
Long text question answering
Text denoising generation
Use Cases
Text summarization
Long document summary
Generate a concise summary from long documents (such as research reports or news articles).
Question answering system
Long context question answering
Answer complex questions based on the content of long documents.
Featured Recommended AI Models
Š 2025AIbase