Long T5 Local Large
Long T5 is a text-to-text Transformer model extended from T5, which supports efficient processing of long sequence inputs and is particularly suitable for text generation tasks.
Downloads 177
Release Time : 4/14/2022
Model Overview
The Long T5 model is a text-to-text Transformer model pre-trained on English corpora. It efficiently processes long sequence inputs through local attention or transient global attention mechanisms and is suitable for tasks such as summary generation and question answering.
Model Features
Efficiently process long sequences
With the local attention or transient global attention mechanism, it can efficiently process long input sequences (up to 16,384 tokens).
Strong text generation ability
It performs excellently in text generation tasks (such as summary generation and question answering).
Model Capabilities
Text generation
Long sequence processing
Summary generation
Question answering
Use Cases
Text generation
Summary generation
Summarize the content of long documents
It can efficiently process long input sequences and generate accurate summaries
Question answering system
Answer questions based on the content of long documents
It can understand the context of long documents and provide accurate answers
Featured Recommended AI Models
Š 2025AIbase