G

Gpt2023

Developed by crumb
A 124M-parameter language model based on the GPT-2 architecture, fine-tuned on 2.23B tokens of diverse data with improved text generation capabilities
Downloads 136
Release Time : 4/30/2023

Model Overview

This is a fine-tuned version of OpenAI's smallest GPT-2 model (124M parameters), trained on data from Common Crawl web pages, ArXiv papers, and GitHub code, optimized for generation quality and temporal awareness

Model Features

Efficient Fine-tuning
Fine-tuned on 2.23B tokens, approaching Chinchilla's optimal pre-training token requirement
Diverse Data
Training data includes web content, academic papers, and code, covering multi-domain knowledge
Temporal Improvements
Compared to the original GPT-2, it has better awareness of recent events like the COVID-19 pandemic
Lightweight Deployment
Can run on an RTX3060 with just 12GB VRAM, suitable for local deployment

Model Capabilities

Text Generation
Language Understanding
Contextual Completion

Use Cases

Content Creation
Article Generation
Generates coherent text paragraphs based on prompts
Example: COVID-19 analysis text generation
Education & Research
Academic Summarization
Generates research summaries based on ArXiv paper data
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase