R

Randeng MegatronT5 770M

Developed by IDEA-CCNL
Chinese version of T5-large model specialized in natural language conversion tasks
Downloads 16
Release Time : 3/2/2022

Model Overview

This is a large-scale Chinese natural language processing model based on the T5 architecture, focusing on natural language conversion tasks.

Model Features

Chinese optimization
Specially optimized for Chinese natural language processing tasks
Large-scale pre-training
Pre-trained on the 180GB Wudao corpus
Efficient training
Trained in 14 days using the Megatron-LM framework with 16 A100 GPUs

Model Capabilities

Text conversion
Natural language understanding
Natural language generation

Use Cases

Text processing
Text summarization
Convert long text into concise summaries
Text rewriting
Rewrite text while preserving original semantics
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase