Llmc Gpt2 774M 150B
L
Llmc Gpt2 774M 150B
Developed by mdouglas
This is a 774M-parameter language model based on the GPT-2 architecture, trained on 150 billion tokens from the FineWeb dataset.
Downloads 18
Release Time : 6/12/2024
Model Overview
This model is an implementation of the GPT-2 architecture, focusing on English text generation and comprehension tasks.
Model Features
Efficient training
Training completed in approximately 6 days on 8 A100 80GB SXM GPUs.
Large-scale data training
Trained on 150 billion tokens from the FineWeb dataset.
Lightweight implementation
Efficient implementation based on llm.c.
Model Capabilities
English text generation
Language understanding
Contextual learning
Use Cases
Text generation
Content creation
Generate articles, stories, or other creative texts
Dialogue systems
Build chatbots or virtual assistants
Education
Language learning
Serve as an auxiliary tool for English learning
Featured Recommended AI Models
Š 2025AIbase