L

Long T5 Tglobal Xl

Developed by google
LongT5 is a Transformer-based text-to-text model specifically designed to handle long sequence inputs, supporting up to 16,384 tokens.
Downloads 336
Release Time : 6/14/2022

Model Overview

LongT5 is an extended version of the T5 model, adopting an encoder-decoder architecture. It efficiently processes long sequence texts through the transient global attention mechanism and is particularly suitable for tasks that require handling long inputs, such as summary generation and question answering.

Model Features

Transient global attention
Efficiently processes long sequences through a sparse attention pattern, supporting up to 16,384 tokens of input.
Text-to-text framework
Adopts a unified text-to-text conversion framework, supporting multiple generation tasks.
Pegasus-based pre-training
Uses a denoising generative pre-training method similar to Pegasus.

Model Capabilities

Long text summary generation
Long text question answering
Text denoising generation

Use Cases

Text summarization
Long document summary
Generate a concise summary from long documents (such as research reports or news articles).
Question answering system
Long context question answering
Answer complex questions based on the content of long documents.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase