G

Gpt2 Wechsel Chinese

Developed by benjamin
A Chinese GPT-2 model trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through effective initialization of subword embeddings.
Downloads 19
Release Time : 3/2/2022

Model Overview

This model is a Chinese language model based on the GPT-2 architecture, migrated from English GPT-2 using the WECHSEL method, suitable for Chinese text generation tasks.

Model Features

Cross-lingual transfer
Effective transfer from English to Chinese using the WECHSEL method, reducing training costs.
Efficient training
Compared to training from scratch, this method can reduce training costs by up to 64 times.
Superior performance
Outperforms GPT-2 models trained from scratch on Chinese text generation tasks.

Model Capabilities

Chinese text generation
Language model transfer

Use Cases

Natural Language Processing
Chinese text generation
Generate coherent Chinese text
Perplexity of 51.97, outperforming GPT-2 models trained from scratch (52.98)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase