Gpt2 Wikitext2
G
Gpt2 Wikitext2
Developed by dnarqq
A language model fine-tuned on the wikitext2 dataset based on the GPT-2 architecture
Downloads 63
Release Time : 11/2/2023
Model Overview
This model is a fine-tuned version of GPT-2 on the wikitext2 dataset, primarily used for text generation tasks.
Model Features
Fine-tuned on GPT-2
Fine-tuned the base GPT-2 model on the wikitext2 dataset
Text Generation Capability
Inherits the powerful text generation capabilities of GPT-2
Model Capabilities
Text Generation
Language Modeling
Use Cases
Text Generation
Content Creation
Can be used to assist in article writing, story creation, etc.
Dialogue Systems
Can serve as a generation component for dialogue systems
Featured Recommended AI Models
Š 2025AIbase