G

Gpt2 Wechsel German

Developed by benjamin
This model is trained using the WECHSEL method, achieving cross-lingual transfer of monolingual language models through efficient initialization of subword embeddings, specifically optimized for German.
Downloads 36
Release Time : 3/2/2022

Model Overview

A GPT-2 German version trained with the WECHSEL method, enabling German text generation capabilities via efficient cross-lingual transfer techniques.

Model Features

Cross-lingual transfer
Uses the WECHSEL method for efficient parameter transfer from English to German, significantly reducing training costs.
Efficient training
Reduces training costs by up to 64 times compared to training from scratch.
Superior performance
Outperforms same-scale models trained from scratch on perplexity metrics.

Model Capabilities

German text generation
Language model transfer

Use Cases

Natural language processing
German text generation
Generates coherent German text
Perplexity 26.8, outperforming models trained from scratch (27.63)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase