G

Gerpt2

Developed by benjamin
GerPT2 is a large German language model based on the GPT2 architecture, trained on the CC-100 and German Wikipedia datasets, outperforming similar German GPT2 models.
Downloads 48
Release Time : 3/2/2022

Model Overview

German version of the GPT2 large model, supporting German text generation tasks with excellent performance in perplexity metrics.

Model Features

Outstanding German Performance
Significantly outperforms the dbmdz/german-gpt2 model in perplexity metrics on the CC-100 and German Wikipedia datasets.
English-to-German Semantic Mapping
Achieves English-to-German word embedding alignment via the generate_aligned_wte.py script, improving generation quality.
Optimized Generation Control
Provides parameter configurations such as bad_words_ids to effectively control the termination conditions of generated text.

Model Capabilities

German Text Generation
Context Understanding
Long Text Generation

Use Cases

Content Creation
German Article Generation
Generates coherent German articles based on prompts.
Generated texts exhibit low perplexity metrics.
Language Research
German Language Model Research
Serves as a baseline model for German NLP research.
Provides benchmark performance superior to similar models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase