B

Bart Base Chinese

Developed by fnlp
A pre-trained asymmetric Transformer model for Chinese understanding and generation, supporting text-to-text generation tasks
Downloads 6,504
Release Time : 3/2/2022

Model Overview

The Chinese BART Base Version is a sequence-to-sequence model based on the Transformer architecture, specifically optimized for Chinese text understanding and generation tasks. Through pre-training, it learns Chinese language representations and can handle various text generation tasks.

Model Features

Optimized Chinese Vocabulary
Uses a new vocabulary of 51,271 terms constructed from training data, completing missing Chinese characters and removing redundant tokens to enhance Chinese processing capabilities
Extended Position Encoding
Maximum position encoding length extended from 512 to 1024, supporting longer text sequences
Incremental Training Optimization
Inherits parameters from the old version through vocabulary alignment, with new parameters randomly initialized followed by 50,000 steps of incremental training

Model Capabilities

Chinese Text Generation
Text Summarization
Text Completion
Question Answering Generation

Use Cases

Text Generation
Capital Recognition
Identify and generate relationships between cities and national capitals
Input 'Beijing is the capital of [MASK]', output 'Beijing is the capital of China'
Text Summarization
Chinese Document Summarization
Generate concise summaries of Chinese documents
Achieved a ROUGE-L score of 61.51 in the CSL summarization task
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase