G

Gerpt2 Large

Developed by benjamin
GerPT2 is the large-scale version of the German GPT2, trained on the CC-100 corpus and German Wikipedia, excelling in German text generation tasks.
Downloads 75
Release Time : 3/2/2022

Model Overview

GerPT2 is a German language model based on the GPT2 architecture, available in both large and small versions, focusing on German text generation tasks.

Model Features

Excellent German Text Generation Capability
Significantly outperforms similar German GPT2 models in perplexity on the CC-100 and German Wikipedia datasets.
Trained on Large-scale German Corpus
Utilizes the entire German data from the CC-100 corpus and German Wikipedia for training.
Optimized Training Strategy
Employs OneCycle learning rate scheduling and AdamW optimizer with weight decay, trained for 2 epochs.

Model Capabilities

German Text Generation
German Text Continuation
German Language Understanding

Use Cases

Content Creation
German Article Generation
Generates coherent German articles based on prompts
High-quality generated text with low perplexity
German Dialogue System
Used for building German chatbots
Education
German Learning Assistance
Generates German learning materials and exercises
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase