T

T5 Efficient Large

Developed by google
T5-Efficient-LARGE is a variant based on Google's T5, optimized for downstream task performance with a deep narrow architecture, featuring 737.7 million parameters.
Downloads 183
Release Time : 3/2/2022

Model Overview

This model is a pretrained checkpoint based on the T5 architecture, prioritizing increased model depth for efficiency, suitable for English NLP tasks and requires fine-tuning before use.

Model Features

Deep Narrow Architecture
Compared to other architectures with similar parameter counts, it prioritizes increased model depth (24 layers) to enhance downstream task efficiency.
Efficient Pretraining
Trained for 524,288 steps on the C4 dataset using masked language modeling objectives with spans.
Parameter Optimization
Full precision requires approximately 2.95GB of memory, while half precision only needs 1.47GB, balancing performance and resource consumption.

Model Capabilities

Text generation
Text summarization
Question answering
Text classification

Use Cases

Text Processing
News Summarization
Condense lengthy articles into concise summaries
Open-domain Question Answering
Answer natural language questions based on given text
Content Classification
Sentiment Analysis
Classify text into positive/negative sentiments
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase