C

Cpt Large

Developed by fnlp
A pre-trained unbalanced Transformer model for Chinese understanding and generation, supporting various natural language processing tasks
Downloads 122
Release Time : 3/2/2022

Model Overview

CPT-Large is a Chinese pre-trained model based on the Transformer architecture, specifically optimized for Chinese text understanding and generation tasks. It adopts an unbalanced encoder-decoder structure, enhancing generation capabilities while maintaining BERT-style encoding abilities.

Model Features

Optimized Chinese Vocabulary
Adopts a new vocabulary of size 51271, supplementing 6800+ missing Chinese characters, removing redundant tokens, and adding English tokens to reduce out-of-vocabulary rates
Extended Position Encoding
Maximum position embeddings extended from 512 to 1024, supporting processing of longer text sequences
Unbalanced Architecture Design
Employs an unbalanced encoder-decoder structure to balance both text understanding and generation capabilities

Model Capabilities

Chinese Text Understanding
Chinese Text Generation
Text Classification
Summarization
Masked Language Modeling Prediction

Use Cases

Text Understanding
Text Classification
Classify Chinese texts, such as news categorization, sentiment analysis, etc.
Achieved 75.97 accuracy on the AFQMC task
Text Generation
Summarization
Automatically generate summaries for Chinese texts
Achieved a ROUGE-L score of 42.08 on the LCSTS dataset
Masked Language Modeling
Predict masked content in texts
Correctly predicted 'Beijing is the capital of China' in examples
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase