C

Cpt Base

Developed by fnlp
Asymmetric pre-trained Transformer model for Chinese comprehension and generation tasks
Downloads 37
Release Time : 3/2/2022

Model Overview

CPT is a pre-trained model specifically designed for Chinese text processing, supporting various tasks such as text generation, classification, and summarization, with an optimized asymmetric Transformer architecture for enhanced Chinese processing.

Model Features

Optimized Chinese Vocabulary
Includes 51,271 lexical items, supplements 6,800+ missing Chinese characters, and removes redundant tokens, significantly reducing the out-of-vocabulary rate.
Long Sequence Support
Positional encoding extended to 1024 tokens, enhancing long-text processing capability.
Asymmetric Architecture
Encoder-decoder structure specifically optimized for Chinese comprehension and generation tasks.

Model Capabilities

Chinese Text Generation
Text Classification
Summarization
Masked Language Modeling
Sequence-to-Sequence Tasks

Use Cases

Text Generation
Automatic Summarization
Generates concise summaries from long texts
Achieves 38.81 ROUGE-L score on LCSTS dataset
Text Comprehension
Semantic Matching
Determines semantic relevance between sentence pairs
Achieves 74.4% accuracy on AFQMC task
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase